var/home/core/zuul-output/0000755000175000017500000000000015157360261014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015157363745015511 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000107546215157363673020305 0ustar corecoreikubelet.log][s8~_˾4m^ RutcMJuA$$aL /ݿ~@ʖ'u1S;=5I,>r`&^$Is״>L[V\Lf.yϜukxlnjVU3ל-&ʲeSwe+ 2(ɒ5$US?Tf דf)1Qx59^t } sq!d.ͻ(t~ZEvruѻpQ||wo>ô U- E M{#3lZ6z|}|7SD.;BKM9S4;XQ|yوŲfVz+ngF\;j{Z/[ #8 &,;π ҊUi[sV #";t)̦ܺ(z݊[)MBQA*&cߠ^Di. GblA}=k5 /bJMyc^pa/nD>8"+~ [R׊6{_LI&%wO>4QOyP+'.?ZӜa*\v-+ŭ5CvE%B, U3_#tڢeN(*Nz]U'1$OJb%OBY5ް;jD:<">ce!Zӱa%E+Y R}<@*}!c#BgB((SVPgt%ңםȳv_O`}/-DI.flƮr֪ViY $Z`}o+ᎣdC-jQ!A]^-]@+lZ2qLlFJPTP!:t~J⏋CP WY_}TH8Ӆ,J%E8T{|L0"TLK?G~`hrF%#;{I(lƄFN Hh=WH%P ]Dŧj 1—.t}ggY)Xo,%Չ ݈|-c"|*2:kCF?ex1D =7xLC/G|W@n&y'Bo%@㢷C~EaX< )AGuel#4C\d>MC?xCGb'1%/l }^w$> ^~ $QY ExD׷'8{w@bB=+6窇Eg.,|]!a]<>P;iH>Xj,]|GR'h<7d@{<HJ:87z|0gyML3 $; tf?>8o؝.ol =!v(B۟MW~'<H_Gd QUѵ'լdq>pTLr7 Pm=scI}RG;UJ8Y-V?Α+! Y8>A.P|u&&\(}6t5n9%Šܝ!Ev[ I߃ɂhĄq d!m^R%솇NYK}F8B& xNn>b!׸Sm@*'N?Fr壴RqGu 2 _"+\>>NTs;2 bˬY@[00W:lmAyRBf'*pfFl\J8>ITTG)i ϻ XRk[p6,,OvYXrTv]S?W']`91lsF Pߍ[(8iuѣ+e\ϺXIh ڗJY8(]Tl`Ge|κ=FfOvF’wLbI/-jrЇ U9pGJE;mF ώLшHC >s?i!eխhVE2 @jVϨM)E)k~?*VȘ[d WSy pL9Zy5q=bs*G-V'[Ui%GpSeΥ9xlj'.+-r9]ï9WSPkhr0 ,~h%5p ѤJ1j-3l4LCσVY<3({' $3#nic|o"Wg.۶i=`jU+5קaj!} ؈'^삿HX,v v)s0:Lsy"q.M_bġ2X})x.'mؐ")a&8ؓN{WڨA8ZCR}+gەg9Md [=f|ul&AlƼ}Vx$oD>>I(}"eKE{FÆF>-m: t < yu`y,7mg ȰUy=z$|JO~Dx;"7 UtKε25#$3YG̓ bC4H%^^vfxG+LZ'H{s޴NWH)'`)_]SU~ϕB 2bP+lYr]1֞Jvx`X8 e Fg0q@vpab.ۺ"6HN!d}M3VW,CFN6܁y|aC\Ơ}4)VSj^z~tf;44hmT]P˜ozN f;UJ~>V0U=ֿ(;Kd縧1tifAK =xO,q*c-ߴ,-c}Ӥ*RUTdX5ptنPsz|cwziי3:iȡ0UUW,Fԟ.]~D1 f㙟qν'9?a1q#S@=1a!3{@Ь"7cy߃2@D"Gn~c -]:ƆEf>+(6_dc+LU92ۘE\XZ담`$KB{(ؤKuy_=5fqoDtPmRN{HvXE=gJ"x} FM`iaqX8StVi@UѡZzvK}[Qr}}xS JYV"Zu}]kyq؀WʇJ>׆#݈h3Z ~ab OSYTkʸFZ\_F.1Y'>xs1~ S8`S+`Uq|A꟯7P)5͍TEJ⠽A7?+\"n 5*ibi Q!cPex9idbw MOW[14)K+Ki\;=8L`[«%/x=BB2i7f%O]׀MXIh0nqNhwxgh}FN'4U* DQ5ɎVV2;ʨ{A9UFM*U}X oV`gLe'rkʗ  9LgV5gW-Oqܼ`'npbp@iz(A\?\S0pg1K5, nJ!BTk!5JGw;ތ4Q"t4(h#(1ty!cj\I#7Q6u wΥ=,4733'Nh,v(&#N&4#kIcD8hyjWBz6ZV[n Ea?S/a#b|iG˵D1s˳;1{u4/Nj&d=,hEpF?]>iohC16Ԣ @&a2^4ң&,MՃ$c-tEaw7$\ZJq|B\s|s5J:nuHGQAG)EO>M-NO\SG XhD1m79PE}T畮͛X\ %6izIaBQ4G&!"ښhmHO9 wDT4!`pD+ S*A{%u$CPtuj{q"Gh]y Vfg‚RX" 7XN>Vh󻏊L4!{GuEOxc | u?ToWc8;^JljT@-߳X{vWB{OM[?d(q#F?Mq\(Xq q |B'aƏm$ p{].H4 '8i_ l6#%;~wfw)^:f͏gffgjִz4Y:C6Za6Wْia~x< ˂ J|ŋ̤Uds)R1"~%KO^xiEY25[^ ‡y\xMVli=ֹ~Vx$=?_ֺM]\?N,cEz^ܭvdnzǔq7{޼Qa2t*57ݲ;K^J?T`8UcGʚqVse-eApW>y4!`a>HdĄ9S GΥ;En 6&,.P!/߾<9{{xz|У_wHxg>SH  qkcA댠pYʂ(f$zpĆ@ēsCrB X;vVD"<ʨ -C/q^oD9 1!)H7'wN`Ȟd]@0}IUٳ$H ',n'+0A{&ht | >k77ƒ "T !v_PXx/!1; ,clC3t,яIeg7Ny( v#gp F0n> ō,&mU^rDO(ȋ[Oaa~xm>H'pyVAi.}Y{ln?/"o#H~_B""(OF ַ#a+aĄEʂGaO ƒaO78pxga,  i, EjG -X.]l>(j#'S7C7DЂ`a) \Y?98|W)Xh`,Chqy,w}=EVP'R(g"eoٗ@)+J%;M*N X}|*x_nϸ6{$LZd>PZdx0[i_ޟO 9FfiAu]N-\]& O 58zSۯ*wo&i?9 d{ GUWH>"9~8pCG "!WE@8N>Myť u|_0^eQ:$+ˑ%)&dys^)J]eH6}\4oѨ>/2K#lo$o{:%F @EM=r)&8/ɤ,fI̧=#gn~HpK@ UnT k,,M B <52oYu9gb<cJixkjUZ!t[354yx =S%1D!T{0d]u5&%FNX=E29H0)md`~؈C+`ޞ)o'ɏ`Oү.G=!T9@ZOMj*ґ4gfHX<6_ÒLS".Ǽt95)ÏmLJA1?"FO>|xu>b 8@ бIy)Givyp>;(Po E?Pl,GA@q!͒A+@MŘ A@X_>ڈ8DGkPq,2C_=Q}(#Jf~zŤ"c.%fg rbv1`ԥ|>\&Kx[vͭBvWp]\*w*`IwQdR1Tgto56u0Tu$nz^#yxM{/e J܌F1"?lo:2lO>Ӣ|u=_LSPY ~rq]45ҩڶՊ5RUq6+k I` ⁔>E{7?\+MHrj>)/z/<_,`,_WY*=ϖؘ?`S~/S5K R]2hZߠ`2A,Hbw p1/'jJ4HE;x,؀JR&gxqyHt $Ag0AfIƔeKc?_֑CeRKH )t#jpƸ`S !GMnZz!hք3\FU!f7IąpοͿ6Sy%,;uЇw1EޮE. s-z.s 8 X@w[ vj _P0/;]P~iմf; Үㅚ[D->C6E?@؆ }\d vNѼc]aq33&$Ġ" $ ,Kq\"QW/xwE$b"" 6BSݩ^7XBw:hm, "X$H[oC7ivD@ލoA䍣mμh""\M[`7-Z8b!J@[۪`mzQ*PhP;e2nHQ62>̼6^HvRQ_륺wjT)I;m0]2F.Tp3xvTGxy]3y` @,-Bx.E^O 9*c0N෺n[PwAu&{OA}jRw6uѨ[սo{vղiڇwz{/0Q{J!CPA SPA zFٸYHy'f<;#[k ;@+U} \9N,3k:T %HfSHs}J\jwyiU;| ϥ>ۂFhݬhtQ'yZg9`|=-uy=OU ,;Bm[zY-3747L6"&AdE4>+ySĔ P1x50;/t4uz 9iK7s>5ʁF om5䍬MޏnPGiΥ~Ac ȹt2v_OLQ-.S=RdW9*"y|;,Q]% ŚAlTpK[۠hkQi% 3Iu=OG!XusZhiaRE{F T`-%|A;@3(v;!".`yLiːC闍i%s ۪ hMdΪ1KayC$Wao 5S$ӳbtRlnUP, %k|l)Mҥ 5 iǍ(?Vu U"Bj֐` lQ!*風e[Av#7`,)mɫv_ecvIQM+,Xt,^Aj䟲b4;l1<̙گ .r;l;溬6x`q[g.e݄G(2Y6X"a8<}NJ1á{fkHwv 6S-U(>1]F+ 9 uM~CRMGdu 64*f,T1BjTU^\1[֭W9俋jUY$) T8:9:Fp_ѮE:rp ;9H+5++&y`dǚa?rP:XΠ #.1CSTGjv[E:X9a⺛͈<6(+Er5DpYxuNTWꝢAB m~:[ T5?Fˏda[N(( LҐv8(^(XOo7uz$@"t+Ir$]ӑ"TV ?Tv5oV"fVi-DS(wHHoQ䞩v:o3ِδC]4=[46^I z4vǵk4v3H6Xm伹 Kkl$yl6 щB$cP'Ǩ\k%4I43X=TpRQ0:4iZϑ$,92N/^HI5?w*U#'ZH"UBwA${ o|E?-Z%)fTIq,zY(7z |YZ_VBMK[\ї]`A Ba*X _&i}/`}}iyxrBo:[՜?λ^^K~9=!K^s9(Qnow|co?b~yi8wL1?MTFO<2VM ?p'i5KK|^Y!Ǒ4P ҸwW|.Ƕ{!0Ufz#2Rh5;EWx莼^}ܔGg<.KqOl") wKg'6ꁯφܐ>#]/l\",7/[_McPRb[ٽj"s2.hRih7&,aN&pL j)! ֭(uzM`hmTT Ŗ**lY~a\K> 64ŋ?ë?~_ giDbpX51iEz4 b/eHoS-Eꌏ2C}&>[<[ܗgUrs\\nhg6||TZ6?fV /MUL u?_φ IGZ-[vp{\g%Ur`FK2Wb(<uHa{?Sn 8pxE"FqbJ'+1aU3|F Ggك]:rm6.\]T~LYaZCEVqZŃنkL/B}LD.ŭySؕNNY$SM+e#?ZB(! 8]? ~Q]Pb[:;J-o9QvрeҊ -HmŴ4)|P;T=>$b.ǘۃ^Lq5 rԶζ¹(ݱ&j.2gwQ2*b.vCgPܡ^XB~lf0"A쁻:`Of_P %/09jK%45 EN `4qipܦP_Ԑg@Yl,銝$d5"8 ̭fӧ{Lս{V7t40.VZ#y2VҢ'Y5ƭ9j^&O 8{⒵PY'j( juE-#/8pK^c`uVh9:J2̎ t>"OʜY|*J0VG0g.1j ǩmm64]SДЌ&YȢ>kE #u1&82z5VK+g*YCP:Aq% ҥ$WcUG;y 53m&>  '"c!@Q)C6`x(pm#}O "r3T'թJ0ys;w!mqL >Af泋 Z G܍4$4CcWYC ®Sf:|ڲ$ =C (ŵ0-5d -L%RI){ D :륆ґb7 BoNڀwu>r*=TJ륂mffV`gImw{ Z5":,X`N& P!|K@Xl:3xSYreґ y,BG.] )搅ݧwo+ʁx#`2p/Xp7V3(Pg͆=X9hbʹ6.SA-ku3 BV2&`Z)q2*ز <$\^Ace ntvBaI{0 $u>3JĭNK|ܚ`HyL2#N+1Xl+35дR66^`9ݡU~/5!I vP}].3\9| BbЄB>GD 92֊AɖRFBF+2qDI7ޡ>EBXa$͉$q^HXʉIt,XZ qt4bBZWZ15p|xe۹x_0E* HF4r %77 kd!KgL`Jș5T&P:__4QʓZjUncZЇ\"GM6EmN,GMQVWCP$mΩŰ~E9:<~*&8vv@ʈ:p j`";QkȌNgDFxpqRQ~yMYFNJixTo(WUiEv*nXmku߽(pP%| Xt%@)ʒce~f c {uocz9Lj !B|"|,na5q2puam7zصmmS徤#UeC;XD 2Aw4QOS~[i2)(zeDjTMPri 20L,JbZlfwnhIej 5i#Dk3&;n4F;Js!Jk#O5^Fvu7~.sB1Й 2B L$Ekb65d̜* +Tb4ўVhCMR/)9Pl գLֹL ;g~exxw(J)vƛ3qFY+V8VP:|lc/ѦK![T >YXS:TmeݿZTI5S0b BSpUMCnwős PnPf'P"F:k(Ԓ}+/[,C.OʁzŬhϗelEeZe%JEEyM hUרԡ(O N*`U!ιNoIqe;]%:wsV[@ $9#縦Q}HbNhVIZ?L |vIn k(~d%evv|V{JPyl֐nPEزTNj0ب(f0XJ}RFWXs`!M{U(ߧ:&8voHzXr 6f% oY蓛w=(Vzs붑z\ޣ ȐAQ45Y q&q7hDK*0|!ڮn' .3?\d6׳˗q)S.tp"E"'AbgYǠEƗᡉeһMYݴIe+im6]4b?V׹9n+ ` 3`2|}Jͫ5&7g[vo;˳0,C;O  =eE&8vO MfA19PtUtloU;ۧSF;5%kOqdjGANj6Ȩ_xf顈\gfꌭ}òU.ybZ YlgRsrH!KZJ)%WQ)d4+([tBy),?Ϊ-!2A-B:A\7RKa<ځk(5O@>*皶 e_\SpuE ;j)g]YBֻ&B2GyJx=[Xxp}@@~t;wwzT~pG~ZlJ\-JFmFIrnF;RixyhZ²$jlU(!Dv^ .5{r; lO/( ]S pܤ`} QRBIu;lݪ.~-%LTwqϪCAO**r:5 dDT}`&8|WB, ? Ӵr H F_N5~Tx}e 6wmdzb&FָgE$@nV0PݦZ=N/Y"WIRh^2vYeQ_ 6sF5yZ#\Sq zMpT9ܔSբO+B8K ˷ NS,U@Gi&ŝ+XûMMeC:gi5hm)%N!<@䞸dm45U u.xUZ&͈l-IPVEJ2"'U6>gBEW*aSoC_ z}84aڢ^Z@ƖӁV33@WS8fVW+.Uѱs|PQ+c(I}¢}S\WWP>U5E8{>¶J}Af5pǒN~Ol͋ArGNsQ' ] BL,&RS1LSnGqnfHcP$6(1#$JyV)lR|bc2IWCl&*fa%`beJ'k:Z4-^hf1Kw42Zp`)3(ĤR`1L\3O+6N tO" #ACw2&'<d) D<ɎQ*e42&Dq}qֳ$ǁgҪd,*86T\1 >BPW¡ùl_" {5=Lg*- )^ca-iHc1WܘXSgy3Җ/A.VϷ`ԨM˩{XVN |;k ) [jKfqF #`.spt)E-J<ՌHj o8p5 kW!HI$B4U:pGFpp Eb`bUd4vE 8y*:^ĵb„L[sĂX0r` 5x>S)^L:Ό.g^{Jn0uI` 8ipZS#t{qUOP0`a=2tƾ/;6OMۏ]%:J*w-R~é 6p)Z>Y[*DɎa׎{ىz_peҾ`WN:@.U$NU:,6|\0%i6 \'Wd*۫N: ،N+,x0@|{׷O0{ V7yX b_R2 VT.乩_3Ld'nT[P\DnY@3j5q+}V\',ܸ8Ұ߬n--MJ8^ĉ.b^Y1/̴Z@YXeT#W~Va躄o;[NaTZ$-isqk-2[fQCQ}f~u>9hiuX8ӛ낈L+fpf( ktYL]!mv~{oqۋS]%<+ނSyE$bN4?*FW棅xc*wkftOzQٓ嬬`7'̙~2M 0e_w߯ڤۛs%^/AnD.'fe,/ϖtHtw#ZAӪ5=1-*Mm`[)HLwk#|~S:S$ , og1~QeVVY1k$X ƀ:n=m/Vږ9H_&gQ*7U'EOdgnc3BQ|Q'hu`v^+vV#! XGJ}VLV[}*y;R$n+?a6_9ӯb{bc.\ejy62NY-QQ|=טM/4rwQ;"0L 켑zQ2Sf7Զp0Df>PN3 aw-Sҕ9%'D@f8UIG'?Oy^.":s? `-swoY-愎ikUtl| I DXYإ^:~KձtL4Q9[d] -(AZPѸ x::zFSc`6p9x"NeƠjT"+9&)R0Н)a^mH$ǀ&_ATHAĈDc֤|\ Řlw)7qxLܠ$v}F)HI Xģ3M \2Pl盗֯ @hИv 1F?c\SK#,8e4r-ӱ'nj G 6.5n"w2&nl#-Wޮ]Y5Yu0RLC!N)L"S.g|F|DM:wF۬}јK'ZM,bN~Yyr =vݻv]~,Ǻ;4}&5fHDaUpw=Hͷ_z%D,I2ĉB Cfߺ|kЯً%$!?3T ̑`-䬰#d6iCc8;5NՑERX"()OXb3Mi8-MvEU8C]#=HŦk=;0!c3(yé903N`Gjt?~v{!s2w`%bjf*E`8P"GSb? ]Rc|N367kRrPYl;Z7{m)qGO[CRU{90eC>+h1U4-5v~*$09棳>:灋1v% ݈G8v¹8B8>f3 cs $La)<IF$gB;$.}ӫ^7>^ BiQ]f'qm!}Ζ[`W#UNnzf5;A0 F;K><cTg;|y͵:Z@*!z"GXb*SĎ ~S(""/ZHe>Ѯr鑔(`x=7* ˻n h$svpk}F 6%>'8JQ2k^zNYRJJzٙ$G_?YѯUU'/]Q*Z0ŗq+oQ]=4X.ӯzax~- ^ ar{AZ +~).Sڭ~f{Qv2rUBU˭Do]15Zw Zr1M+Zu,En8ʓ2\*ݤ#$}^鴚>?6#4J< b- 'YR&*aQji  WIFj,R-NKgh*3v\9&8YA3j\%3e2We[ISĕ2 빞s='4g`PG"]ΰ$a$ VIgZi @QJ7S 2:w@[@}'}'A9 ?QlreJD&["OoOs}bJ&m[X"560ch GZ7eѤLh鼳@ ?cfLƓ9iXƸ_]V5 qzhȸb? p'??Mw鏿;k]$rOT|z5n+E~yc~705O@9~.@/gp݇jU_T a~8Gs+f_[ @zeϘ>@n$B tsu>]L  5~^kq{oGO؏SSKusZw@Ob~͖gLKPj`k]W7kܽItׂ98r)"Oo"din)cp!BN,\pQ#?]SZXތ~,d GrzX+ @InˮP4^kW9Ź6qRS!I0"$~ E EY"4F+Cr~nB7qϥ^9\fcs4h]2Hd-)sua/C YwqW_ǭeHjp[6}|wg4yl FO x^jpJ&Jˉ,]T1Ahk#j={!"UAK֮P9J7&:@܈R+ДF.efg-f4Y*>݋W_{mX`3Xny 2Npw(D'|||=B"|6d)[C+j :|Η=b7!ZcVy_! !ʫuҒ1N ib}Y~GU|ɘq7gz4]CIFJh*j.=%P pDZ #T@PBe& (8-'%Key7|#=9 i gɄ$GUU瘤rJ:ƿ"ҢQ dଌv)4E\#3LN&b~kIjw>Z#<"SNd[|Xf m/9dK-LcRoF6GS]#L+-i1tae{E QP2]&ޚѽ6_h`tamae^:tQƀ(; TsjHca$M0b(JS0lP֛Y["[3Ē^qBJVJ =v2ÈꎃEr쌃nJ\$* z[As"F#bNBJr 71k)ŽK6FhZ)XP_ަ,(HO-X|2CoOWK~sEw+igv;AJҁeIE4lTIEKJj+[6bTm .C% %\JK%:!d!8F@DHa,h [I,Zt:e2А}RhStLjS@WcT`"ڢ*2娔”K2 ]3H j\-C*βDBH0ΠH.lEjwІ{9ʱqQ; Z?;>PSNw 0`?[ ~H2ź /CD`̢ !]H )KZS;^oթ.YdFHL *SQȂBlǂcJ-x%X;RkȪ^>X5V10:*9%NC2ܡ(uCSVUx/qZқGam[.4f;H-D\));\.7ap[0rky|G|p6Q.T9)K%bRJ'La!B)UyRL#$NY+vY\Um7#W{*m>N"7"0?}ٖHpnG'+LvDhG##\v"#kLpw\_B#xg,K$WXrzD!oh,L , ݙYɅwC$,4T PGdCa]";qK[>0î ]Q;_ ʇvd,eQrȔD_:E0mHkQԎ%Q:;:<5":DTI!k:cW"mnW0с$k^b0V庀bwQ;zzSCQ^?-mm/t^JmT\%ۈKۈ*ۨՇwd/n;ɢմVCп#`;#״lwZ0xhY{՚- 4om(%Ng}}>ɝk{U٥L]Ⱦxb9X^?Z~mOOOO~,d/R8MCÝenؒs6. ?7N,[#=qBɸPm~oħ쥅}PA$-/ױ>U۱YľH 6rMkhaCG a z@ pIp@ Ozm?.g\G7s!{>N^iL'TvB ¯kǓw# EQs'|\Kx2Rv%ŮFy1>LzԂC;N>O/B'[Х>0ZG{y+FEE@LemQގZpo΁VPkfޭj*N|W" 'IOOs os; ?\mǾ5wl 8z[&kЮf}"]ȍ Tf\8*I 8!뢲QR\ Z=5zׅOIH0 :=iF ꯷ vQj=- g6e &'v4-?i `5pUbySy_aG)j+w^aǫu^ety+9]{fĦu4feWQKcgvI3*B_yKؘ{5uV-iEu~wGR'UT%%g̞A`U1/'BlU ӄf?JcTl &,=*~wE3]1U>ڽbnttƆ/-7;r|1=`7tew?#-:zZ\zAr_nQV𒆿&Tu @.8T8z!!;RcPj4*ӌ2gQFdJq-Gvm ydi@&U`HfH~Tz=.׋ѻ%BOJW*XR~F=*eTAJ~ZcOѺ(hXCz͑A-iIO/=ŸʅeL>D7JWZJY?ڗ |ZPҗc_FFDJ6D*`) ґH5 ,QtJ%G-]<6!X&# ,ޣq1.ԃz2J#J(52{7b#2 |dhH-\/{A4JtHnuKbz=2Ќ^kudz,HM/32Ll.#bzY"e#PZz3 (jFg&8+Yz$cǒS|F2epGW1Js~gem:_n_[um~8%rӮy<[*յ1l:u mb} ۮ|9lV?tc~Ui/;}_nN~77waΫ]|G+黮5o`{ Y^:DUwTF:sg{)n{w:jgmNoϯ-¯[]uFS.t󝱆箫עz;?oez8D[.$ۊCNVj$oa: C) sq=^@wWTn ;U׳B[Zrh&k-f*"$\HPB|8=duM:}ol2GmW'M<]*`jvu4D'dj*ChV@9nquۺGoř}lkooA#D鿔F P +wV6m)|^!gϥgA Ay!r0ER$ɇؒ1چJǒKf>+ eg]l ::괤Aܬ:9ꔀm6H5a ـDG ZPD53.Eߡ%K#"t=IɓAQ.Gq) ;r0"@1#@rPrrL[r9s?WwxM/l{>8$y:ݔ9@+_RQ?0-hfH_.1X뒄<$!ǑRD2.z=i(`7C\}pw~{su|_Kp+ЅY bFMTC_sq<#aTA&b02-%5efKLۛMKj`}'d q-O, eı9eFT l'#}TF^Qte`$BxC$cvDBLɆbRy܇:4B-$ n  U3``Yk(` kJ|PLDֳ+(%$/ "@ ٰ$qd%븗’O8ow7VlOONl_}1xkBy1v]_9;Bۛ^u8 M<%Ada(0 $ʉ)y^%WIR7,{z 8BJn8,8)+!H.0rd綑SCU;)sۖxY^׻N[u([| &eq9͐zHLBl c2iet m߅2ie7%Nc7٢:F$@MP'#Dj|Yo[10Dph&,ڗ_n~ "sե1~ɒǚʯMHۯեOckbҶxWYhCu/xpGvuԳEx=)C 1t=[ePC oSf5,@xr3C7Il dc%=p(y05TTUes57р;>./ A FCR+|?_F{W\ ET#TIdlhHNx O=BU'S's\$~SVI)# A9#RFGGk GS -/2ܤs)lkL Q9E 3<3AtȱOT|(ӭWi!"]][uu*^n`&8o7W.Ë(A!`B!* fСi JāYAnjLuRSִ w|#F$LMS ^[{9֥(h&, È\ՉO0Aޏ+5WjJ]?;DWMJ*{o1&c*H6ٯXX7'mSGYxUm0u]N crč_|᳠Xq !gA`5$X1z;LڶLK6I5,g@j$ Ȝ^;~yK~yX7xAsaMVi'O'-H6ze%p`]Tsh}ޡx! :d- NZlxR>^taTP ֞.;Owu ޢOdZ$B H 52UedNLUTgN`8!qR1$BxCes-ZO-qDށ @٦@c:H6p1hlF1oƞy$x NK ظn&-i*i1\hЮkyfw5>VqĐ)Ld,5@ț7]ljٷ;e9 85cT™B9n[bM16pU-``h[C_*:*Tt>|I*H!NwOtl(2DV؊/%DVs9n0Rf$!zKeuPT\0;"z5qβxRw yq`w@Yڕ$TMU!`%g`e|i)4J|P9;XZ=㪍$Ad2H6a䲙::zƾTξڍXsƒł%0!H>` \HB`B"6/A P$B련Z*ر kz]7$e((Jgn"HPP5$Txlewݙ>yss}yCeŃԳsM@Mŕuފ3- 81m֢߆3ƺhdצnS3$8ˈwF4Pоfĭǚ<7swaHJC PV"~Pj{#3V 40bd` +T,XҖ%X\E "hN&|2H>*`骼UGE *O9"'^lšOʧ"r|"oSy.T7p`/?0,"L4$8pLi[O7=x)逾h%/gD1CVgC@攕ܺ|qcJN3 9DJ~s.`@A KN&!w{ϦxX+6" ѰOd|d-I#R D Ås-9O0)( +nk'Ӎ,FgTcv>`DdkR<$ {G%zߛ?}Yw?BA=2A|7 ,AKrWa[[:shbͰV[ldit`t5;2*UӉsxtl4-l#iD',iޮ{zʒ~>o7T{$[B@W!WT=֮7qAlDcա ,pe:H6\qhɖ-=k W4=Q@yQXm%LM =ؿMt,aE@% z2H6&hd:_gA+D3uD\t|Яx|E@O?9{dz&wvs?[|~uxo\ߴR $پ.Q'Ƹ/EkCS8Db f|% gJ MȆG '価6*(ye!h㴞*$+"R<$)S6C j^84A/g۶D÷YED+ƨnDN6*^ʎEwdKЌᾩTXB|MjsW%DlM&Z]Fv=2 'YWUO/wj{;ow+ɉ:-\wmH1w%4{l1v/;!-Dze4/ʱv˭g]%Tvp^%@ZD=/9eI4WNt=f=&īUhU =qJ H-> w#/6҃ f{E5݀1G!XzIg`=\: r4uԅ0u7 F4hGiGrHRD;v?Ò֛L EqYؾh>k#aկWnQd/~]?3Rcqk! oiD}Qxh"}c#|a" jD {`>SdCK h]Ҽh}p*̣p`P*[z" <֟zC'k8{ #I)hN1~ ːiWi_h/PťRC)QS_(J "ʴ+@W4+jE2ʈ(sJ>(D^uLn<)"R<R)ɾ,Ha>t .^9h|J0o&.}s\`J'$_HϧO"kbJ|+BבA(`6Ǚy|"hiM6ti.L]NTEصJ: e3~>}>8P>%Xɓɇju*H"!ُ88C\ñ7[pd{W/e^4eDg +W6,.H"e* \J}AA-vm8ەGX9_zwwځUΛ8) պ.?m.(xLgS[-ЖɸZ FjmslКFp,`C^-Pr9m2>Z P9OaI$Xn43Efc| !9F'٦B=n=nOsj(F(“c;B]!cBi=@{F ոDa=\%ޅ6j"J=(G6 c/0 |DjP2]>z!&4c. *SF%ҏ1j BD ]`l]t6Pba?EzҺN)%g.9{32.e[%mx{7M #u{֥|m{AKe Z+1Uۭ }˿^ԀՒy8aL3Q˒ˢTdT $ Z4R!H!rmdz0@OG,U^OYPieۥo_o_lY/97g${}xgB6h7 S0q}TA2f(PuлV y{L-58Juv+'0бR=VJTm,)dzGYCP{pZCtlsL`\4Iŋ<,A5h0^yJH hAu̼# CU8m1(ډqGXUN;z?@ӨbEYzrDhga0M)txq :i*s"mah=\5yՑ =~li,1ڎ62@IS` @[ fhm@h%k6u6v4Dl.fmYΔ%c4(&ˏZ5dtԶ jGWmqI"ukAU=|<gUEb?uY[ò~ [wnwYk;%5V9ۨK;Yk|s_۠U1.ALiݻBl{f028 2y I 7-XF)q e'ch_8mS5Yk2jBAY?j̠mR{α3e[unHi=tnm=c+P0L)hY~-0reUdajЇCHѲbE/y[w#ܑP?)*g"Gm,Ù+KdkbfhۃSlP׽>*c Rq8UC%~\(ؽ=ánj <9-j^t֮:gx}jiZ$.-BV?Ϩ?[={G.Tb-g)'Etpzn\Ecu}T_L?9f/MVgt?B-ܡֻWSeY]/kvI]vloi4j^n}6\ۏKsA*<_݈HOBap ܇~qB;uwYr^l:lW[DCTЭDvw?&/npp>uoTl[7-rB$Uw%wI֛M]Y! 8+㔕R䐖JXdʠ? @bI :%9aq` e5(Ɓޚ`lN#J8FUB%X?h&Cv`zc}*8tMA0G9Xvqق,D +#u+U \Er II7(*TA*EDdR 5ZZ\%?JX^gA+ ,Knvу _j.rੲAJ{PHyVcB8I$dʦ$rڠW9ͷY%䱪RPbUjVP steğ dQ0[Xq bXx?X2>h(&"=TGC(!p9G1 FT[p Z" 99ZoEAP9AI!4< 5,(TW(DAAN ,4@  G%e?Xs&8=vJ׊^; =P@\ylg2?(CG" gf3 + %VЀ2mPEMxk4X,ܶ2K"jN#p U=XT -5Z"$/jM F5ʚOGT]Q$S:J3PE5кtA^Bt4+.8CJδ rJ#aY$H%"ђCL͂'HA\ڥ/s@Ňͳh2 YTSml)K(u6, P?uqG|&倃(D0#2 M%\X2/ y=M3Sj;xTS1,b1;^)ByI J{@ 9RG-9[fh2p-#EHF#{;~4hwC4 t5HPZTFϑ '0W?G`C w NhʹZ_A/QP(;ҭO(ED/EHŤAP&Fƨ˂ Z'BI,rHVbQdRW4K@JBC1~.:Ҳuayj)saGL.AЁBqN+%qwCIG ZqYK]l%Y+<ԄEWzo!MH^EC9kp׊JQWN1\}Aޠz)vQՄe+K! ? Zn`jj(I[/.B撕Wf yTDbKaTIR$Cr1+QKjzs [T#♻GriB J3eյ@%EH H4а?d%⺛Xs?X(JaemF#UM3@5}E[f(2ns2@ MO]垆`12-~ij(<шDפ`5Th#-@4pI""*+ 4IUBXHP`r%!ۯ$p&v̈́~wYwqr&?G zQyȃ#%eunYg~{oqҾh(7Bbc~Ry\]nvieJ7({R8?),lv۾4?zFw3U_>3Ng-Wk|z}:=9AI>fC3)yBGWYlENW=5s7S YW1?;uͩ{tHNWͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:ͩӜ:9u@ƭS V8uhOƩCO^SG)=:u W$oNiNiNiNiNiNiNiNiNiNiNiNiNiNiNiNiNiNiNiNiNiNiNyNKcJNl:N$\'Z!_S(nNߣSǁ 4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns4Ns"é';4^O~Nx¬ Wy~s~ \J$\w3ߗu7Q}Vf \o|A[åYmX ukkKcM>~"`1͉Rj*a@M!&䶩ΖOJzhFk b|"`9=dKTZVzF.UŘJƘJ60{yO-TxAH̶?[\\,2A_@Ά Fg8u9#'3ܸ= 3&d %+Vl*`5VL%f 3Oex"`}n1*X'X P˙L`Ua"`%MeG:J}l*` O% ,RuN犛QbUqeͳfQMF=[D*4~*pTZ &T:Aq"`n܇ 3>C+N,-Ҳ̮f"`P 컏`-C7'䳰^tv }{~G?MP>"vO] 6Ҩ?D ~]B|b2z6o¹ R}A$ѻ_׷vo~|y5K[˴ɦ*SUXZV,DC⨵ƕȓ&G|m)DJC; |VA\WzEbm?{;?%\'=20@ \PF#Bni8u{J\r:8K.ڙa7n<oOtWy<9˟)aY|XT"1J!3URǂ,1a2/XIen?Wމ= (&;\86`{l`JJZ7y*1PbTZVKΤX#|"`m+X2ļԌ$)Vע6rd-z\'բ0çNA^J=%+iqRLJ+Vh+)#U21Z&Дs9V:N"Oz'!gC_gߛ\Eȗ}PnnmnԐџx}b55 @2;1o_tdѨ]BXOfa|\<hp~u"J;^8SU6DWҥ݋+j#~Drz?Yz $Ew l矑߳R>q3o>&~Z!9rM]`%/7gF᧙&YU0`,yAϱ`KmHmOZlǹdY"tjuK!slcO4\)v[g*II>d+!q.ED6wɑ1p[N&B!O,RRsKhg+?4G_Npy?ݜ/})íA9L$Gru6j-.ZF53F}|~ƅȽE5W_\Aٖ-%f`V5<z)k?9-jA|d6<eXrDOx\ŎTbImt tEv1s2{),gS}CޝR-izQݵjbٶnp%N-JZ26ܡwn7# UR4l*AZq&SaMDjpKiɽ[_Nio ,h"LYh*H L))wuC0N~ UGc81GYo nskҋe9{9XVG' AW|޾AM&z!$\8Бt( ڨ'blkV/rL lGDLdP'3 ~ :y[Z==:e0|BT kcz}D5Ri gݡ}9qHjn3*9tD1u$$oqdLq8H1PW*J %"N%񾅒[P%e ~k5b!an0_eKwH9'z.= GZƢi:oٔŨM>! 57d4.w렘s({cBƦh|iP jZu4;9:bf?rk4HMlkbEP4nU$0IU`@5Ρ75mdUC9;0LPNl!,꞊jf4Q8*Lp)/ !!gM]tEՏVy-]@JKkVLUHej`QK `JGJpZ.P[(C2X]gY=vh,c!.(hDC-bni8g9!e{o$e#ӤU:| ЍiE86Ɠ!!CA-MtBksB"C AV V^m 8+CFA2:vRBK`, O/T+κkCDKt0)( &(k:P5s CU?FkcW MYޓhVSb:Gz u&"+rJ*n !%^eU!4IJ}ҩ(wJ憲vZv<&i;aDb2hM ߣ,BP l>$/?0r("Τ*C/HZ(J0 2?%~ƀAAI"K0/`bb`CU>FK=ZaW) E.Κz4X+-W1 `ɪ[l-`HY s,Q @H5#Lj`5!ލ5ڛ)#Jt50rBCդ j+8;x APJ{w>X6[={\ PwV&b>ޫm]߿ywwIY!0Y}j 8=p_O$rM!u;. uwvzZ<=}*o>V_N]_G7/+plݯOz//T:o" ZK+\9MNmOřfLrW @ֻcx,MDԈjקoֽ͕^<8I?ooK쟧W.+Ͷ9~~W _cFn}FF7#}`_~ _ǚ *kpio?|⳥}f"%ouȔ~po_dCxO:4P^vU]|~kz{!>[2-GPU~;VJU|뙖uM730rrq}n{V>؛-įdޓ6ndWc4b IȇLdCe˔Z ($m;]H٢e !)$nneF*{(VkjqnqWiQܾAq˔m[ d)RX0x0T/WҡRldKJc9ҷ#KNDŽ6L |˥q ՊztQ-k<$:MkX%+%юj^#KK+?v,Y.CW\ZBx9j\JWtW#)KbF\R{(JKWz`(ސJ=bib.Nc7Й]"A{JiPlWsdئtojz+UQCU( XX*:5!^PIىr6Xj 1Rl[՚Lp@XPҢ9 sַjD-('CRG$B*f½߬WӏWZWMV8LJ?:˓@T" [%8 ̴(]< Ot '@ӆ"hCRxk=4\'~a UNfT\Ob`?~pU7U)=:9pr@mkw[JEY:`RF&YMؕ١&ݪ/</{6PiVgI)ߡr >2㭯BXNP L \|6Cs‘n`Dnj=G{o9,(I+UXݰ*ZA oi=kdo{LWH:̓6;6" ]gG1 3-}i30Pkxz94 S'ai`,2"-h|S[T *}S:F+tx5a#ϓBy01esUsLӳW$5>= N>b5MQJQp!@fRMA\XІ{V9R2;d`W{Q l'TY'<0ev+ }H{/0 װn!XXPNQ(u+/TRK0dAssńuC˺ Lj~qRe2 -E5mB],yY}b3LozcTdX aDf-\ Ɲ]Sw*dVbϢ!g+j8ѻ:D4\F < eȥ-]L$'LLBbxRRRP``4`l}{9t`G_;p/B^s<ߟ]?X0NZ{ V{Evx%TlaΙ2οR*K[MBW>?y?hLe&{Vuܿdq0(;$sgg95㭽اslбYk3@D?vI[6 !,p\[.Ce{PDJkk;J133>EX%lKVS K@Yrma*Xcu}6)!dZD4SBu$M%seD^UH+HVnz.[۠RA-wl! !6*[c &"Yc0)gc9Pg#L% #|}`>/lDJe|l4&hG".B-b [CGub#hwF (AXD$<_Q2/f0XYX٪*{O1lp޳'8v?MU]FU?e>d0#^>J-#M0hTڤ^%x\Hds{}(8}\m(I9vnNxG#b^oOZQVh%GgyJt=Gn6mrI)̽Y ]Zs]n\r;"o~@d?~T PrH5C%'GFZߖ2Ѝdø"e5`*;@m {ŲM/,.<_ ^l2JM3'sЏdVx߻}o12vnQ;zRayQ. }N,}ojc6OsU)+)3ytV- 'q2xj.kRҾƝ-o2D;XIcq$Pk|-ȇ66u{aݛ̗ZIUAxIQ\e0}g!޽ٖ[WH.KA7HJTz')ҕU#r+)eh0]k&Cf e׹%3P<:RM57^#i-I^(yQ&X߽yQP`s_JGEa+l25z Hqi#rKŜ:\wWJdvg5KUrO"w'뙿 W˱g!Â>2㛄>%r 'ͨ"G θpX9u:[=ݵt95ɾ ̭O_/^Ȫ.0My<7zMnK`pU3LI|3e򍠜B)_y(2~G']-$GØd.&  ߙ֏}O_dp{jjւ7a:􇁊Sb1eYÙ#cy>ܲGW33=314-W=$mzzqEu6ܗX\3_dKm2Nt/ %I-(6/Fw=.K߱aGה޿1٥n ~qtx-_Nw\͞r?3~.)QW#B8یwqH`f,4 A)R3)OF䨻:ox0Q3y?鲍 N. 8nIf*F/̌^ݙQAӝPqÁ§x1@[m|MOӇ^詄J)aB溅duARLx3&l! !e#}z oJ_^)ХOˮ>̑]{z>NiK F(= ' -C!PJP+}청Hr|Խ*{KF vZzU9b Tǰ//v5{I T0 1z1 sHK `x,Y5w*\ͭYnJ g ,ŵɛk'|/^U/.V Q~Ƽz_Vנ懼t5AŨAQ?eZLGB ^W4[T%@#1L<|1i܎R?n?JY^c~Dάr煢s_}O:u &YRk?ۋߝ_Q\rq+Nʏ\Slw|gw\_]=3Y'|OJImΫY+t./5ȨI%{de4}kUUÙ <_Nv];mv=z{FnȷMqRfJ$jjݕ@?ug.p=g}mPQdݘJpƏ[ OY 8F93z# 흎2HnӄJtSS uzZA-hp )8gK&pɈPc9h#qz];yx9ҭ*M\:ǬF*ҝ Oҍ؂҂I*hkٯ&XnUm&O.#'.v+n1Vd<}{hk%TlFac{'Lr=Ap=j)3!Y4( nCm=}dS@O{jpAS~)A< 2~f p+ Nh‘9\CdAavm5"yp\GFRFA!ebD $.% -UkiV98vK=ŠUɧ1= Xɨ`8G #ydڊ:_9=3\!uT6ac!O!p0D4'x{N X{!dƍonƣLepQd^ä~n {_6 (v̲TQO!kYIN{zIDH2YB@dF[q,c MV9{ҎNu?Q~4]aXwhve(ι ]f>OT=K"LB2fKzIeَ9*>U2Iލ~1[P{%'o%to_-+y>O}(ڛx 3{8Q\r1 LZ2YXXAL%HTx8ңEY&s5`}1MI-?6(PΔW ړHA3A}zm4.ѽ,r3Z ʘI.z(Zm%޼ bl=0i2-DOF(oCŴ#=U7gOH|޻ r&Ak9 ҵQmd7?j3|iKO>壞361T^Keʯ}'?/ŧf9,9ZӜ 0:dAsr[o=L̅ejhfJB Y1eLfF&Cmmp6^9WG7Rp?ڸ;ųoPƨu69#h>FnsCPR]5j%6>덠qT-JZ5co_6˹Oa_o Yemѕڟblۯ]ʃF9<ԓ{bzdWuAٵgM_=^NB½^TuuBm`t67MMG ]YsgN#9I-$o.b3XWmz EƳ I|>j1F@a;e$wp :Hic*G˛o}Ticri>ŹLᮣZgs 9a;<<>Ypӌ3:sK0Y51J?}jM΀M0Ies md6>g =7mzYrm`u7 ӏCi~PӔ{P%}ȋRLh2\XeH9p!gCzij ߁"EzOùk;Hɷ37NɘomfG'^aY: =x87QF?#xծ4"GKu_'Dj>SX@-k:d0HOvv H;t\IAmD|DmIdmfmG׻dn乏%=Gvs;8#T`q*y5js,5)Zξ@?>Xotj6FAֈ};x#KUkw!JtdU'N̚X=wI)h|CR6Ok7@ҍ;1yB%Qȳ`X<Ұ,+bJ 綖PMdueV?ӌ:^:m &`R5k8QI9ЗvKX58:[7?UJQQ_z?us-at&,;J-u]ad($IW{TKrL ԯ2-[/{L0;L`h<-Dw/Wŵ|6w5RL?D>b"D5njh6}Q^h8z̠F)wF[Bb`zX7_hFe}Q· NnO +D_ ^A`cpΙ "2D:tY$??), ?aGfxz5{ kCXr_RzF+iaKIYއFf@Z)m6:S~.Y:sl]+>ghƟ.wFl Y^Ո.ՈUSqgGhsfFQs/?Q{>w]UblP8ܘyrNiSU=b<\Hy;**%<>Sgy-ob[E\r~!0鿯OC~D%hQgD d&9Jh,"y5ҿpF>L~/g{8W!!{ߏ!%!`+ A ӢDk>VT C.ۻ|+VڙUml |5 ?D>Ki}Vrc`hG_b#\ cvU-ڠ_OxVy=b ^?MkGؾ|> ~[8|o4hfZ8N_m@7Hɼ7b )mFNJYH~&u5\PwFZ*BkƙU2L=\SnZ'+4kSUwmurjHsx Y4 5a6  WNjhko~_QB.F ||ǃ9yᣝ BεP`kre㜜xxqtkmƶz-Vy6G*+-hx.FLݖA_c/Mir!?xwkM`}vn(_<-VX_ǣZ0b)Άw,kی)Ѐz4tx T](zuT /Ǡc;ʳ*x4e"?Ւ}`:矯j];g[˒T4lPu?rc`_No eFк`o7y9iLyPOzj^(Z@^+ Sq̎'F,]qo 1G6W!r~~/godG5|9XG|y1 6deCT$}Nnes]uiɥB$ *.TA9N|SHhs!%5Qco콮6(-6R*R G(!-q#ynɣ8۞ZvlKj9SQOåxN*eh-CCyZ8uӒ*s}PtIv8TFTG_Rۡ!Z;w9I;X/;hjO_akưh.&|+=?jVN R`x<@dCZcHHzl[8N.#ѫX7q&=XGr!EMÆ:ʢ klbS"- u" aqpDD"P z5+X9=h5Zf/0.6>?2>f~ޟˊFl]S0_VOL:K@mKbe/.}tb<#z㆝YOnO7Eڼ+Ϊݿxq']A5O ʾgjTSf_$iK;o0ҼAv=+⇙/2~sr>TG[ݍaVm.偔El+~ /9?>r봺ycj.$ͼstܣuSfUEYN ?WD+B!mQ5{(w-\͹EW/n/y@Xvoܛ7QbPO(A) /5ɷ/') ^V~+p giܯP^~_m8nmpfZv9 YWr)=X\\:m$WM_|Ӏ+xP*xqƟ#%|6MnQjHd?mBnO»Uz,Ѫ]7GV@eP3i޼B|;#tOӶyr_G-yI;W. }(UcM9 .AXJ'i騭$$h쨤 evLHe\Ga"nt̯]fy\x`\ ;8}ā]%x< JcgbdpvB$zݒs :~hݫ(/mi-Chq_o;_0/LJBR~sRFP1dTTI p!z΅؇D}3nR-ݩ3ޔavvN SG.ICI;rq90:o1q8ɧVގ:o&u8~L;z7 O< _zni%Ov[JnI8+pK r]hv};5oVUiۤms]j=)mg-/o8LdAIB *8z/%SIs5sAs1,9\sjcu'zlDr;^w/ p;f1u"X~ey{~=sz"LHyl>VЬ^?^tI!̮f#`an}cdo!Ĝ a 2D!S s-čF:x,&Fmbb=fZZoA7p̛w z\׼*ھ*rsaŸkrZ+s?k@W"RdI!eBV% .frC``wTLjt$qgi:3Ŭ0igrźcvrmćWexmal<`RcK<蓷6%88Pl;M `-KҺnW &c"`SG1 u.a[ HHִ|wiOͤThq{[GO=[ ?%E5/J]g }vZmA벭ۋر/jY/#Y  4X)9sppB$FIzT(Vj_SơN$\8"uۦX^).C |Zw`r<־Tf\M S՛c|"#x XiR̂" %ZT;z1505лXbr.&&|oyI)X4p\2JYXĨfF*' < I%&/ -gPLQ}dxƳom'D֎55ݎ?W|N.9htcZeFmsT@8A Wok#H\̗<ȿUNH:?΁EJW<%+>k-R&[4F/\uGe{Yۑ|j"i;,,$d =c'Lp"ZWR%LWS+tv%%MK舼|_K}L]%QqJ 9$TETsQ˵_uqb5d~rM.$=e32QS˵Hdz'Ə`L $E\saJI0)c)+rMN&rs7ŔsCS\D" yw?"'sv1<?ͧ6Zh€s,F45|< m+} |_aKyTv>6e+BD#.Jp@69Cr gtA+hMS'KD,@(*Q]("StgtxƃjE*k 8SFGwK"?a@C1_MB \#AJaE@=e,E^£#2}E~`!8)MW010ғXS,, n}x+~qH#&_=daxۡB2 ZF`3-aV"NN9"F[+]mJ+b~{yhX4h~Q]#:}O&,z:8€)pQ04o K|OQ3|:ӧ@ɤ@4QY'":t y-1(!W' T?~Y]]o sZy)a5`MW~78T1 @˚A].Z[S(\m/7J4WھvfPsߎ@Wq#ațvvKk{WcWwR9t=R>ecRoo+Xq_꡶!cz1KvQm/襪/O`r:UqŶ> _62N؉ܓ=6gMgO=N0n$Y}5Kd?R)6隞+)"⾞۬kepSh;gĚ˒ھ,".j|Qi~}-)"Ez`q:=`s&aӏ%XD;?nKY ӗsʚq{5Ġ\/Õ·n\HՂb5(hr.ɔCﺫ 9smǑݸ_K\~Y q%b~H6{Fg4ݳlj]zZ#*]*֥)c Gw|`n!*_| Z^l(-NZ wbEو<ǃN[ѹbGý Lb^: M.:Za*O,iF!% L8.BtItId8t63gay?\RL&nmwqhz 6peĢ1݇pU3]V{\!N?:ֹ*.wǀcT&l&.VOnaQ%C"6*g='* j0s훻{PXOOeX*˜`]U%UMvR,↥XԷT@))K/5%+ JY\KVTSH@NTkm~@c.) m-Ǹ6:9oXa^vc~l4T(8lV._CňNܰ|f/5QpMnd:%z[92ޅ~]"/-J#mH dY?mb%ڥCqW}yo9JjJ$tMMVFafΧfڶU]omU N7M6pRFάuNⱋ5zH;so.57\X)')Rk#'Waǝ-$v\߂ UnUS$ e!LU[!*Iw$,0t3+2|*oMi:_G-(M?);vQ@ {x nd(8m{fszMR|wozD e|09 2A;͵_nhtw/GOe9 %\ +0!G .!E*z^ԍK?=F4kͪpuxNdE|3z(ߪ4_R}1@*꒒A=7 |R5RIi\`JT u]Kw%*YyjڄvgPw<@*n\?}n9ɉQpn|Φ( 7t]vX<_Ej3LRcULN+(z#Nl_Y#&΀iч}̈+_[=Nnf dETWN1ѹ}w>咗gH: A:a{|O6#Ն%~%08*b\[UfhƑ='>:>#I,jϴl`]#1" U du>B= !l6Ml)R#~ {2W=mU\ *]+Y%-*S,,>{6Oeb%V9]kl+<XѕckA*z oژr2@W QkLmfvlNXj:aR#(mIenkL*} FǍo|D"x͍f\k=Ծ x݄C.mB ASIڊ#' VĊCwV5H*%fF: z? 狊$ )Qn]d];Dnu@uD1]l1E~(<{hj.Pd+wW'G@%_l,N`9ir[G|b g9ȴQyͳx#eArnu|O|L:^XD-;\ Jƫ̍Idҏq%B'NU:;P*%w/ճk H `v}'bU6(q'i'xHr'ҏ`%eF킆S>b@t%M.-NFB>sr`bGq_v{s:=d>shAT_eh'$ZG(kkO.E3ءlȖag CN !^"Լ.@ @`>HB $-d ^P/>h^Ͻ +[Ahu&R[G7oӛ(2ok„ Fm<""g $ z5[m$ JC$zcjq4 Al@3 0+d0H1;(d C'SL?P~Is!z)JT~g'<z #wC|G@!Bwa܅_ިF|F.ј [_[0'P[m<@7jLX}̮|C̫w9w?! @,Mc3J8!$"E.inPբfp\&I{z 6pGIh$ r.6s4/} GMnN 5ymHShН^^O_aG|EyօJ? dxA~e$;h6pyQNd0b SMsFj%xȃCnZZCcXA&{{@|Jތ&j]TB.i7yP9ߞDpu?XU2EKAUDHw8>!C<~Lw#EvvP8BvA!͸\ nW |Tj' ~~E+A2qz\ g%T0R\lm?z^mwk۾lͷCm摣$BKsG TcUUF0{f֔Ӭޗ~)v|4rQL% "> ֳ?֟⒨ٿV_ZVG>n~eEםr=Z}]mVtݑW)f͢ϳj3 ٲP>PDT֜͗E4Kgzai4[)I8oa0 P.b8;ZnL.oPd-Kx^3kOF Ru[c䈔TZu?n82pT|S_75WK&Ʒ58\BgGQFuA+@9r9kF2v^^| $rQE ^踫"XFTgKbD;q7"X -r!* i54vv ,O.4PD˜場k r[|]j77Y4J))f@녍xawj#>P`|Fu>\2BiCJhtJAY޷oc?tWm 1IR1@ĸ,}>"H~)ZeuSI7lJʦgMGx;rW&YRxl=Ax^!KD ,!K|*w)o0O\amf1z\?p6|8]umK9بA.mYȕTN/IȢSaS /&l}97Iu# # ASepݷ3#Ն&zmj%0);hlv>&L775?Y6LFi1{/o<:ht=G߽OdHx}}AoO6du?pfA6陧_di&%tͫu:Uj(n.|hea"aҸ(hM[4fUa4Q2~ >ۑ*ݬ߫%p{uɩ^Jb!P*1R"7qIt i޽DP+cmry-З /ݗmCąƪM3rHdn6ޤ"&7PAn09{#F~1K~4 m6p&]\ھbJgKGaXۛނߍ9.nn~16F:6po62H*%ҞME +ďt7j`zYklľE{g 3'2bZM*1|*Gq/6HE.di05LԾu+>,jrI:@raY.H>o#@CchtoTtơ~k!\/w A+y:Zu~~xv9L&x$+nTF]:eytFYo\1e52MiLSUuc%Bp f*co~*tJONP7RܽO;AO_p&)QW\y\mLK'ځ' rn +( tW` Iܵ% IAUVq7< '8N݄B/pk?^{^ } ܎yD^K [͗:XV@#a;D ~%A\u/oO:0xDF?*ti*@!D"x͍f\kw&#`udOp z}6>3@T!H9 A*㹼Ve">α1,T7@S}rq1QUr^졏\S< Ch^f3*_[oonC6m,KNWh؎HkRQiQfyFbNGCpXuO:RnƼBX|Y4WMY!˫DoJyiYPG O!"%f-3Z/$\7}h\+""[ Y `n!V:w֫.bg+qj[z콒$ƹ&ǃ ^Π0Gaӣ0xHl}ظ,y"T`*L*OPFg9%OhЎ t߷K +*p-0!A /wm_]Mr%i xm_qC[+EZIp֖6ZZEKںj5C~HgÙwm{ 2y,V+$*W <64)) y>`V(tȓt8+~)2|9RdFa6Pܚ#tf(rdW,_#9ңߍ&_}/WE@w7]KŝJ# j2N E&P2<|'TLӫKw1XiA 1LCC6XdBL'tfvƒz=K*}!pfx"R$cGF>!.(f F(LnE:]u᱖ 49}!xa`!L^+@Q4P uX[ 4@ 9V-obFoEau7/-iò8_4fϟ5ݩՊgyβcɢp/PMZFƤ@J@Dn6sIyC + Z>wpo_?/Vr^iyA1%"䂪f 0Pmj2NEge>{~*EbӢX70/5-Yĩ+?]W^#_(5a|RnuRw;PY+>k%f=$_ք|[xbAO3۲c407>?#(6h6C9+H)=Hʥ2c ,XN+CԖ;pN 6*xǙ5PGɿ[qà{yL.r0iJS_Hek0 .ǔGUi`_S cY&TKB=i37'β8K|R&^'˙|Q kmMb^1M\^fWjG~3_]eQ^n4_zt/h"2r~e6 ZLcaV1,˵کqCy{^f-<=`籏E`{ohO/˼{;Ǐ'}gy\Gqt(&Y}^竘uLyUŖ57 T=9%U*f>ܪ,p2h1yRx+kCY2MN4Ȏcv[ Т.fN匮ǤЕ¾m°6w"f~˻ CMu-&qju]\Dw~+*ϫ&,Wwu l0غkͦ{;t}H1^GG1&wy'?|/vc׍2˫%`r^$)Y=|q5c_>gwq9tlկlړ%,!I-bT&è 5mhAH z_a SͶ M;XwYUŨcI#Dm:OcSoۮ8f0T>Z6QX<(ƾ<ϓ |F!w9Ji#sJf ,LϹ118h3<D;Ƅ_mV[ǯ>È&^j2~m(zFryFh~F6]˷~; ,61G{Cb;9"Jg:n[윔(sBk mͤy _!1bzKS;lÉ>=rA忺CKWuNH=\1o6kO'qZWփ\RzH i<=T¼/8mlN߸WYLGuV7Jp0ElSҍTV$.FN[f P]gB Ey$ h`e^Dcwhk|meWeis4!-IN37.{\(Mt+^)Xsƺ @/g`qb!#C'fXs =l*X;U{jSKr%HƜ349ihMu /[\xI 7 dzdQJr+E<޵꺬Dz䒁@P^pD&Lc"@Vmȓ^r0|]8zPDKG~8bLy6,V}mr$0'纉bf&HQR`$j)4Z/䒞Tꡔ|ھxߘܖK W{. Y(VikrkҨn}yR\H o2#2tfN=dHc$h+ I yNgr:F=_+[KQ 7.1N5a)NYt|e,)JmCRJb-꧛`%NJܕ@+xIf`f=\`eI䉰i)8r2wv$ixGHz}O%9^*%>GB0gr2geg&Pö&hɤᓲSr v؋H)S9ߙHr@JPl e Y!I8UIV֤U v}QDk=a]&r6̵_ڔ2m§[!UNy<Y:[-2ɆMS=zCSRn>$ِ;-C9@,sx@ݤ9p!4'+;IJëT3uT7UGuȤ'0; ˝bw'qa#wI83^FsF#PJN4Ip8B" 簳ll1QaI9A!0|3Њ қŸZ̑voFr/B}.E}1`^? 0m)n:x zT>Q`X9AI029g@B( VT٠8 |k>vARrAgXײRt#2;gF5Ł;.?+Ε=f#c-2PսDX18X2`ˠTX0Oc2AP1ʂfK1,87իyi'QqI~ABK=O(!\H;ab#Z)E x#Mc`pxvGjXҤv<:hH?S*RL 7'J^{k`$;9F=2h9زac<3FQV*tM~`H?^9bS [<=,! ,LBkp1"=܄}` ؀\`uH9&<]s1088{P>Q2Kc)XF>SGHu9v|O/oQ,48+\-G0Ca9E:qЎv\ċ-VC%2tZ$W0w('ZaqJؠ(vpI}}<"2pDI%|̑;D6Io}1,8HU'cR.݁A Ƿ E[`͘:9b 3#-!S(`2lf2Cڗ&ZΠFg~4) K"p'۲˸eHʧz(`!ZmJdzR`/8ə֢;0gOZiRίÂUv{dQͨC:D;sTN$l+qfǙ =*r^o6q_aNiH,q^leMaZBK#9ۑX ɲGb%)]%f(wm#I_!~vp` M+)G$߯DdYM"HbQ" C_,7UDQ֣>N<8G j(-gN<)M BU@4 F" ? x7qjJ3UfJs1m ]9MV~k(nN;J^_]7Jw۔$Q{ J>N<8/E>ԌQ${yE!F¾*(W.B#aD"˸IbfFU'\8in('& J3Jm[qxQ2gꑳj2"N?Y%}IywdtA3{(|0`+m>iKo "MF0wگJy᪜0b_گT//xQ؁>N<8'(C2DE>N88QȨ66mQKnx"*axzI.`>m=ِdCQ(žtb7`a,}0q"H,sYd,(gu}e7^$+Wyn郝/QpQ+2S{~Az5_y_4\D9U8v;@iZQCbĵ d'﬙=ۡ 8%EЈ\G7>\g1eYy{ָ^ijP.0m7r\?2¡fK_ ϯL@v"BǢ)8* T+0 @O/Pܫ22/p(TpΝCPC3re+j[rƠܙ5o<06+K0C+ÔRk_*+ m8bZoqD8.`ʓ/^t6: 9L-r(!^;/i\?6Y#V"zJ zkOxA6)Wbi!DA(~C%+}$ޗ^QTyRJlykգi:DA!ī׹ ˗v|/u+_?Ic.d>K¶f]A `~<=rHJq*-l>Cwbǐ]VS8 Q{ _KٖL"Ce{\Ln4U0g3އp⨐$Ih֙yv Vwawx F8l/ @15R bD$)':"Bas}p[PgTWi ܵ;KGYU. 0TI-w CK |QvTtݷk P*_ jNz{ƅsb7!N/J£̉;Fx3Q [I? #2jȄ *J2MP) 7I|jFA׉2bP1,cVk1BzJݿx_Y!~'` uޏ:4s>da<3KёrnTӏ7h ϲ4н$:10$g4z:< ye\=MCmiQI4ӱ.t"7'y+~*LDDbS"NDRɌcww uvEiU3(<dZ]g+&]Ƶ]`q6r@$3!qE`Wyѽ)ӛ| to'IVN#*SBVUP,Yٲ`We:WS};w=aMKU`4h X]moeE91esn Jt 3 -,^6VBjS&7٭a7. IF˵W-MY/(/CCAW_K8@zEÅk(<()[M8A݂?J+PTrAH+U[}Id/=&)3.)3x=R~Tgk;0i!T,VݩI8vhGḌC[U#f*];$AҒPjBFDl) I␰(gZ%]O *[||0a1|y?t|{Q7nkRm?N3+uTF QwQE7*tHHM_ %R1#(}}pbFogcD|LbQ/S2>=0+uǐcE-*!pR&t6I. GN\9,ѭ`ˆ"{CG2 MJJRC1 p kJL]p2ʈʘJ퐸 O.5(gp@JHiV5viE#  `!k! ݣz\)a#'8ݝ|bֹ$,gݚps=5& 3eSϹi#NCKxdDOC,u(fo~` WZ~RS{/31B:3Ƒy9ƿVdKǁw|}p? _j ?gr[:c?=0 ϛHagy L֍JŗoCQ"#iFogcDd(?j>yZ —MPYF:?p`;.N:TJ#RלԎ$KsKe.ϟ}ԍ;_c϶1ζH,fMJr?N]52е8b*<?QlU$:E/ǚ;c9}nƗj^u1֜B'kRFd &!Hpa"$9(T,6 >Cg;Z/,g2+T:MySFEe**AwaOѮk1ְP:es0#IʙM2m3cї}Hԍ_gc1)؛CU$w@5Yd,XPz,u@5hajB8F9\p4UyR4"Έ1%-QpAa?.,Ζ!>˻G-T"GJq(䦉 ҽ SdK;5QZ[ήR&(c0+즸U?13{ɫvh#!IK)cDhdBxc$e ԔdHNt ~Qt*?xQ*$g+d`IcJ=A3[."-LdQLôA}Inb8\Ȍuq~q]IP~ =EJy~aYIG#%I͋whAm*s&#qԺʙ2M| ,6ֻ YF0{6-p{u) 03s;L[^! uM0'}ꪌ{]oꏪ_UD7+|V46gXE"7?5zuΜZЫԗe[Iu0VOz }]W]|,""C$ -6O&}zis1bS=\z !TGX|UF4%8*qI%ӧoh^EEUj6Fqc*܈K |Ƃ aNxJqS"Dd4X2.C|nXܻfPvwTS{o;FD%YFARW/)1IXfLFa0w&0)q-J~YxRڞGI{邊H#>YM=x;ZD'r+m+C #LiH. }yTe KcΈ؉aHXi0 .egퟭč [gfBƅғbd'&{+h ߭P>}VrUdU>%o^P񞥾qg Vh{Fn,cއŷ$y b ikFo|0r]|}GY~x4awj8udN׃_6M8e1ݫDKr7)ߌ]擟:]VcQiZ#o&}o0QZQ‹Ij0KUsM!081$%sV[zz1\eyqԀϻճqD(T4+`hv|@ OoWDe1Dv6Uo0#FqvOf'ك<μ<erKOcʌ}F11.&5ݽ}b링6eaTsINixmw2=޻z:vO0L&1z5qcA"ŸI[FHjAEf|RmqqG:qPFC ˍf\k}j9VJ/N2cX!dKX yD&5u /w/"+$0dM7J5vxMuk:NͰM?]ժ}%]3'.8Q՝y総@Zu8T#3О9с Վ+M wn>O\̸;JȡN"-=$ ;WgV:tF=NpwoS6jEH`UH8e^9t&d,?}ߝ3wVWGȳ ;Wqd#t2ǁ\'PO1}$5>YQL#.A%XDfb[`U:b 坓[($-i+)駟g |{8z±6撎n=+ɵ{?! |ПI ʜfHꓜ`j$8DT/񾕌F]|7NzҲS+sw $E L 6H&oɽ;j7? b [b pqz8*j7{۔Alj;xwN+M]L!p ]`R?tgPRk )i*""K 8@eHP#qN:45n3ĭj:_`dLќzѬ@b\Da)5(BSZKLSL&&@T[LЎcE1-D f#kW>v{*v~K)4i[qۣr%5Qc;^,#opG(m,_8gYSKk@2?I/1ާ*-S_#-3)(1 uGҟ{^5н;}v/ YQŜaT:q$Iq%yZ~^8 [9KKwYyvTKm[}e ʁݕ[.=8Y\Y߷ GKtS}ɂz '#auǮ &)b(r !#ERYQ0>phXR4l_t*G /~u&NF1Y4-+>ƙ!au+uJ %Qd-HK#1` MW@)ńLȷJa GiLSc}L۽st[GN]9//9r4ݗ"I#s6cb/t[d)'a8hIAIh"GarBVZZ9lVw,P;;=*EGUEaq G@wX }F&GeA`__s2@4"b@P״d`Qc:]4M jK G4a1 mP. .י p=VHQJP6.'Gwq\twzj_/Ey&CSLPQ'N )܁ fɍf\뭖ӝ#ֵwn{p#Gup+ +` sjs=)o]Q=sts:rx;O9܋ӭѻE>?#Dجxj-lYQj]1rFid>f/N@jYH %(kD#ǠB{9RMG )\Hd$زUX?E۷>cgcZpM?(BjnsOHC?$3D!,EF%$ &% a-xs}+w&=T~܏FeId&)$6JER0#b1r8mGA,9xwȳڏYRPLi)Đ+( ZV/0RӧTXņQ'"I,!)O"&&fC rP홣c0Fe|teAZ IUHcQ`hLDNB!^$ vY@ YLYXFj?nybA9Jn `X8_e`wv9:F}g%1 #γWi NN9"F{9:#t GDvx2pry0gӁ}cKv#{twoPJrT1rgj[Z@X=*% !0r~@"znѥZk/*I`g25E?m!xŭ?vv((J8גfo(ɶ)eSrMw6! 3C`ԤVWgDyױaw;D~˓[y5)EI-)^GV =1 Od4S=@IŨ(]M3;/G%wzE[38w/#ِ^{ :9Ns`%$$1ϥRfF*vdJYX`')8dذnG ]C]hgyXEekFf8ւel*4W:Agddx;,e iXзI+JeRXE ~ o;;D@}yv}:bPe=O7z4.̭.`CX&k3[H"oVD'aCڴ291ic0 3'MX1ح)0H 0Q8ʓ&Ā׻Sw߲:*wq|\v= YaO#wQqz>[X+pbb6pɞS'.z!ø z4}NvNo+u PrS Ф!db(ƉW a΋ž% `d\*Ŏ:*uB&b7*vX-{Ζϑ |x{+hTC&(ئ2g1αZN#mF`2LNP/NPD˫H׽n{FAJ:{TKP6m(\#K- f :{߫0@ƪ(?tn38$ TypУTzʠROԏu,w۲Ϡ~uI8$LN$SR8Ia/-3&+54z'@T;wDDR.fK"$q<^QJ=c |]{+{||gdz6E\_&5wQB>3.4H9b:ny `}+@:|JIX;\'חڧ#  Sw\'¾ wThPcyv$ !ǹho$ v &q`$bak WX\S3> A $#yq|G{Lm F{4wpR=-顽[4:D(ɪw@^qzq>KUHd5h8qA="ܾ^Qj}CTY#n wZdѬǣCg  _Cw6/:$z!F>|ʞ0){Ӓܑ)|=çT\w'uh/:;h iuז[`a 39$& vN63J1*ɅI*ָ)]X(NghhC^[4z'@|10w`q"3v+!S611+fDjFgJZ8-׋q(bw]kΏ.FSH-؝GKWF=!v'݅ԇ!t1)kPNX=>Dlakóۢ@s`Ei@e `aO?Śj7Zi65щ ^>?lir2F @Ow`pmT hyC'>^EWk`ph13mV 0;q?۟j RWI$_;evד?~ @MW,Fg1Y22W3/3Ź$XJg(lbl3#ǺuWʻ]|55?[mL_z^~/M5_x9x5Tzh3k'_j['4Ej>Mud1|f'Es' n0zso"~] zx [m`UlE7M@pvѿf;~]ޮ[]L'[^Џrmwf;dVPrUgcIuQIx^@WzfC&VѼ4$5iT\ ?W#ԛ @&Vr~̶S㷅soWv c7gl$Q>Rc7Lc2$pj@PLZJp-8'J*XL B|| 7z f[̱}9~gE-B%Ke-Z5mVwN.lO+ҴX~pDoӜqb]Frf%c¶tjtV{uv6?M5|4[̋эx>*܅}|1i66-Y|e;}foq %˞Ӻ<׍+L\ a2kJ)$d#B2(ӌ ,EX7N?-9')t֭iQƛ-q$9T}4ۇԴh.yh?2FZwh6)>6tYbufӰ! hR. P ڧAxofYT]88wZ`Eg k*Djk]ui(_U<8x9NFGqU)a~ЊT^4}W;3 hX?>U_R=YF";ҵl^GCI 8G 0[s5ř]]S !tظu_ &娼;]^(^]G@"@Ge9+ Ε [ .g(i12EG`&MlV1NAV fu`+!F|+" nfϓfUcկq V: _-H#Nw)f]%FT攠4¦+fh!rSpc(P0e #Y%b\JLټ.Ly|41V1ѡΣZGX `Va'YV4Ҷa` Jqkmt4*ĤE R .&4֌SHG{M'AU(buytnD]6wnmVDa?l:&ZI˕9<T.^/| ;?w ׀ܿzzXA+Q燍AP 匀U'˲4SI݀%b_v*>wFٲAW7^J]_|¥N_BUL U!؁ AƩP22qw!2b;U7f9]0 [`Kćqٻɴ(GYqAoIR3Esy]p yDjbu:w2E(m-Bܨx2~[/gĆԙ?^L@n,npGc]#0yK-9X{aeWTsReF^=W⇽;׻}M]Y{uRhm-0.m@J7гhQGԭhOvC~X /kXOyb]6,W YwqO9b@{QYxhF?N15"^Jl6V!17E*)RSE$b- EߟGzږ̖+?In,PkuYZY>lX:TVj[i >cm}f͠mh*nCЉ }06m]A`gt3}sW$w|^!بtOat U7H) q5 m>D* .֝>Eplb>BJa޺#488m2AU;@'c!Qwe}ՖiYŲZ{Y;N_6-mڃ[rP|P- %012#i]*z_6&D)(B7\ńJQ*g Ժaʔd.$t:/a]u֗bw[=U1}E`yOX2'O()Q9mũ46VRɓ,v?g7?/*Suh6h6UK(t)]hՋqj_Kwz]G={Hj^u.i_3 bƀ?Gq) f 1 _8= 8X I_H, SnK aUޖ8#]1CK/u1Yмe^TY EY QhM:{({!Xq(gG፲c5`5 Cw,c9:IU7'Y!8 B) !($֕PT rXddGTVT:/IV֒=)149 M]?l KgeL;uQutn]G4m!Y;f3pHQ b`n ̹lȤ`Y5-OOX~k&u(,9 d)QBeˡ9794x!&BlB/rڤ9#,LB2f˴s$2bHpD}&FF3ό偡tB"SB hc$E:MR&cI i,v֚1 x Z Z e6 LX /ȤI,>?LZ"?n>z'1e+SIfI|%ȝHEey1c6LL0c2 D;pN (,p,hLØ*& E>`Eb ,9es']T,cf,*d"`^[ 75QGFe1bliGϹ`P;T[?r,i|ymEB2.uW:of)of}c=h.e8fOargKۤs8iJfdwΆL&W6M% <1X!*4e #0 H %cq+(_F sY`V͔rH@#'-#sdG5.(@;v'ѹrueͳ` {Ga c(h*lt0%U[NAR{cB@toz : +¨^$c.9PIC p!43Sx8*%DzE8V+ZAi }գQ$ɻg3MB"m}f6jUŲ"E|rS˅aF U s0i-Yqmឝ4"ԥY}QEx$=V_B|qd]CkLw1EW(Nt`{w53YΓOtj}Z0R!0-H [J߽/B /WycGad|U^ήmol UdNU}ePΖe$ᚨ /l.Mo`\*{ X^c,뭌*+>Y!Ry]x1^V^.kw;O4̵qY-Ջ]2^&%3xyi kl-]75Ce^,o@B> GJF"h8-auepV j=ȦV1tjҷFR\};T5(~{2ӵt<ﲲ|s%lǿ^~U_z?_HSe@.up~9 Ё;M=Vm5 MShtjkՄѮmݻÖXEac. b/n^ 7t/{&%Ǽ[D%뙡6uH\UEѫo{_Iw]MT &›ΪvB𲢵U]za=X$ _ ce{ /*M|v18N fDB<RVDWg_&BS܆Nr*a<:U5c+Za-Yc &V&v:'".+_ p{ zI L06pqǤTk4&pL\8樲ȱBwWL} $䫒c4arѩ^{]oA!wO^Fx3Kv^Oqyyҽ mW(b w2j@VCqL1E\H(Wl=F}Cd܄eX pR֜iCRz/ð7rI4yχ=VN1q:v4-}6./pd&Fe ]eh^‹f I2ޟC@V~Y~~|ϋݾ“ٰJƕnul:Wz"+ hb5[vT-/΄.ow\;P?O):>Tk̾p(>'cC>|N1SgܥBEtO\ N0hP@^Iƌ>;o3Yp:mC m|9lOd>xHlpXSlaLa7F|&G&Iػ"`ys ,L̍u&复 XK漿 e»B,uHqLt4[jc+U9&S{MJqlk+j}%] YPYlSڲgؑ^C-!U3UJ˼!h@++o=FS}C cmu 'iNFt :s M([sZSi;}?=@n9Lȳ}>W,W5辋3g>3l .X͈H*Gu9OO~x5.OF79U*"P擘ÞXxnKb} 2J Dc%(gm>eKkK m:XX$F[t9ٚ/?^M(2 vt9Rq&E >.o`,'Ylqָї R9x@I/\ZeR`m(NiK3Ig73VEuJ<+~~ _ܑx#ú'ߟnok{X/%]:Y=XAR+= :TG^ԯSg祊I@,Hļ]eK UN|ZQ") iJxJfyQuʫWV~^HxQ7P/TWHdz0/y94{Req=( t?{ރػ&6O*wn*caJrh#4Iu䞅U"\$MP`CSr/^QE&by^1e~֍tⶱL 5=P9[Z,2/RP!@`p9"ܷAkpkO(m:Kwli)]uyw㠟O;P \H])@^;-E0A{dE'K#xɭqRrƴv|98"L@H F j P։OyVHUX#kX:h31H&xB8HS-S@1H-^ eaA|)dY5-;J(cRISb~.P`Q{E_U3MǵԵ4kRiv/\8H6sp3fa˫?m\VIHfVSd*`) #B֡C+ FuAIW wb.>N6]0k@Un;"K- ڕxwm6!Ɵl1"o1jeI1q2 $*JT;86>n>"-L7į.(dva"z2Imxz>~ kZnMB~?' futӤ'pq> @ iK9R*;9e6V^MG68K]6e;@avf5B #OL PGw`jJ6N:u/6liɖm<.tUn"G*L;.un+V!$Z/qCRbIjh >rFJ|,qoQ%N kb|Ge7z'mU#wP47LPUjD= 196QE$vKC8,i*- ԉ6IpjDQywpw튓 N$T;$ 2stJ.u%I !F;#|dA\FĠS9&L"["PJIpq<;'te֜VVEIs(z%z;DxbyG">AHrF'8>D;]|i҆E mh<H 0Pu,HLsSf9JrIHN9V.5OC4;Zz6%nAzgb@D7RdeBTjgDRGcy0{;sQcuy̒vݵL\Opn]ڦ - *ΤnBUyt'_`hX?^t ]iV9i3>J}& .FțQФPghtf&pp1.~V]Pp_쫙Uy$ƴox(flLZ@DgdmӾ·5nq(/njiDVj.@[V*MmgmJإ?Rѿ Q);D_Lᔪ<)TΟRT#H²7:|u6+Bbg?b;8%hl&;؝,pnೆ/z Njƽ?R'@k|AnP]7tT`&^_BK('N:yYfmސTE.Hzj/JKe72Xh>[:Yzw^ Zޫ-n] ^  92a|\p[,5d 3*l2U9fuLꉛsW: v1S h7==y|=sRO74!T1=9K_0^g,5sI-|k0,HXr^jqլmr~ wFΛ@uKnQ7n ʰIn8nvTIc@t xhBZg4e'xOɶ -{g.E|KhC:rS{V)G.L4 & #ZVz!BD4>T], w%i!&CMH~Zs2iBđL{]A.,?  枀9l>Q;f4v)+% 7(Zryr`P 興 yޒ TG~8أUE5Q0yJQ0ED٢0`:Q<+Fa.LWQ`7.>+A?6! ,ʌ#"EZZsvX/,6feܨG$0-hΑuV>L#%a :84!hLٱFjȇCCf!| ;(dy5&/w蹖2d'/ #Ha&^>{z- WHQz9'(p!ES&br!zFt K҅⎇`z쭃$nn#hemtoFtD6.A J>qe)Ǔj.@CQqNĬ^eҹznxjLjrܲU}!Wgx Hm"[JGzRMb٠h@Za=3jƷ'F,T# u%8糊<iC.s/{a--1>&xsL>׎ !LUx𶙍,o?t9F8Si5X3ޫbw up|1;cPQ#w Y]-eV?N& j1oN/oÖ;[a].C> aˤo.8G]`=4|;}nfoʐjg+Z7co)UH.GUdaPK{P[X̆i%5;M=xXJ4{]rtZh/=CK~^̀}V /ڟ_^|*|ms: (Y\koYh/Ú#JLV.tҰn[f %MPZ4_i΂}ϖ xq2P?#|$h.}v]GۏfKn&q[Uz9xmju; p~\KpaKWMJ:u(|{~6 k|" ^4MDa__W:@[[i aJ+N37XD@J͡0BٜPiKbUf5٣٪496MV b*ȽNů}~hl9qoK;i`.x55T|iѯ6psYi=NF8aԹMM'XVoLMf/ 2MmL41@Q)-Jeɾ MհNWUSPi9NȼBI&P7 isY"x<)sYRQctɀqo@\\6kKDtէYRyJQnq~ᄋ-oL]S{g0{{HMwz6~մ=PyygZ#omdoowxO{[~v曃Xaboht{o07͓Pw8|7ܖwa9ss{i2㟈ӏS _Y0u_>f;Dz]7˫:Wb+fD`ֲ$Ѝ2N41slPs s/y( *gkj!8be|֜Kk1*cͱBU졦:@$`YC%dAFZ dy$xi:."7 iC@n]򲇜r9w] /t,%zM:.ALiD ugjŀ[bV%ҭt+&Mjd[RloR-?;!+9)6sI{BU=NU\Ngi^ճR3)f2\]U}~r(:ւ3& kz˴SUb&2#oO=z}ŖV ^yvw׫I8rtO$1$5fgK8K.-̅H 1JI;PX(]| hCE4i͡p9(" lQjL\B!vI[d1u'FE_;/ =4 @x^>0ơ3w 泒i60SȓHҳ=A Xd=((pNV i! @(ѫEYCLV1o mEuhejpeY{FndǼ_Cv2, /'0x5#K%xݒeY-=MUdY,3(}iw\9bK7?D;_"mCLv>U URŃ?r$tn8hjQ}\G?Vbɜ(N|0 Mfs"Ϊ: S v\Qf9<]t?1HP{*60Z U4?-C%ȳ`$ߛ0F/'F`^"v~CeQ\eqw1~&Z-~0'uj۷5$G(3Pϧ{!oB-4|Zt/AC[[^s!Զ';/͛gw}`8|K``fnە5>Ҏ;o9]#i9Gvڍì˯B!Ƶ̳ hZ1W R}]v*l94ӌ  g0Ʃ{g 9{sS 4{`Լ`r ~|zN>wh bA 8>{CZ[ 4chJbNJQq_ueZf w`ڤ.K;rZͭkjɂ6q= UX|6jʷ]PVNP q*k}S? #VQv@=uS1XhRq^,gi[ sS_<2KRjٺlkLuM^O;4INKgKSr&JHtKu ̤LΕMw jiw+\| gJ},hC#{CeVB<AL0>em3GGlQ$bgѲpŨvx%qϑ 9q;eP?)jdŜ-+֮?@ ēpSdgƉR0WHM7"$C̀%s.ntǔ}T^ݔL{0fBuwMpI{u 99:"WDYn6$R[SIBkƙe-s5ԞQk]4YaK8_] X,>JndTAU?:0}^e~β+^iʡ=Y\͹גɧ7Y^)C ]ŇQPl8N$x$rPNF{pl < :=ՓXς)T^ťNR3Uy/|akP0E CcIJ82>lxWN0)<:];k> UkW/㨚A>ibczSUwutI7k{+ʰoڇ]'KV.H)>W..0@h_;$ Ӌ$۬{;l&?|BN2L R$Ck @)+ 'L+Lhq1gxaވ, c1rG kC ۻu/:"t^ut_y ^Ǽ7/mOQVOYa/A}7^E.o+b &ǟFh5ٷdhG Tާ%~ĕb9'E-i IҸޞɗ`O]Z-MoYG11/\V,_}HNȺﮢS٫_G|L2eQc7^;JSZY} wSg`KNB_iHNgo4mr\{f=3'v:+dp9= 8 sz4$]͇lh ]aP8"@\%):qFR=G;CѢrmT"^dW FIrZ!vU7_ENi!n`k?VJb8'"FYeot$Ǚv&ub:*[)K\Dl?[蒓4(s~yt:"E ,1O:DXGr$*0I)g9ˡƖN5v 6WGqFѮ=eVvV.w;zx P_..&)Fs!2/0[LF[5< l)Aao%! "ൗzړef| #޼Cﻓ_=xy(;<Lg hjd>:üӨ$X2o5XG[Y`^ gE^Zqg\p`*ʥFja6v T2hmSbib0ភ{!z{A<"@`nҦ()vHPe%@}i=;\:=sSrIs0r?ts_U-R2giPW% ,o׶ySWvrsIMؽNIc#Gڸ/+Jd.F0?]m`3.?~8"VSшklD!Ͱ`A:L2*Sm$NNUyNajUZNUy3l YKZ1uQ jE)y_fC(@A֑z Nji(U0ƒ$pDiҌ꾲Op>}ڟ}r o2~㋋,(lvaR{vT4 k[틛ztqA>Á~Dh=>WW7Gu "z8g3:w:&y?M)9@mf-h @Dڡ/{=cfV=C;g']Ajj']@=H6n70,mKȸm] 4qK+2K|ۖDrWz $eԤTӀ0ӯp/jZu^wh6ntA6̶EX4wLPj7yH^vm~F HVCe@ uXw`)k*( D 뤷k}riEMcBW_h Š@th5#WeeLXic$1E4%q\cC.yfpKL 8O6Ƚ׾]߶Cd|B0wo2ʚ#M~Y 9׷07/vvI\k'-)iQҎ^4jL>#>_wBхXғT*c$ + e0> yQp㛥v%Н LT#k8* ~ƣӍ}s: B7C"R>1^m*D`A$_:3= =CQZ$ @}w_F> h?`2OI $B~~d)%q,i1iaMȇ,>\\NJa$Ϸ'h\"E1$:$Mara\ V̏R!H)4~)8=FX%JklI"\9 "[8N"_wjB=>[fA9Lu69UMY]`q Q'zrYmX-iDwqbJ$pP9;^:G:L pցf^B dHu1 =z(ȇS>=JF:Ys ҬK*C %x) AdH)dRXv G#/10< #nDV3ydJmG~ oTbk{a:&thy ȤLH%h> R(.t/9\긊_\?'HX+haDty} M8p`cMfU-1Dj0!"<;F3N1r\P>ZԄrhكzk^5VB Nx<3Z)bL> 2ĥWq `0,^l>x2Xelel/cwا?3l/jk1}xq s >cJ`UzWH {dW(/ ю=z(='*@tm3č)#èׂځ=@^B #O<R9KHd@8ۑߧkTi)ńLqr8Q.̻<0Hqr{䒍 3휷ۑߧsNK<#jJ`dS6i#aJ /4G0*T2APy$.Jj'JcDj!c}z(ۧ ,K JD0N-(*",|L"hF5꿢tMpgٳEN`đ\II6|aٖi[Nv7 $p[Э[ gXH^=o-G0GЂyP^5;:xq44J:.$ P4SŸ$.oM7[AiOw\^]ҡ o4B%D\t$ (A'U(ܠ&*ԶiSykBUI7z",rM0Y/݈w0ݏyN 6]Hl=6\,.d)F;L4uBH- %y։MtC=s.hͤ.hDF]J|FN ri+v;Pͅdx݅ۍ1̓%fDzj-6fB!$ @AB;B,2 AaM3?B mt&C )1zE+PZ0oi}'ȣe1t¢+>Phuuu %xEdpXV,C(`m\.^'"3ԕ%2PhΗtN'I1HWc ^O Ѐy-l} /.2+Q3]gQ&!~Phrԁ`ihGP8L5=b?B t@1V-3 (X63B0s;B A*՜\b}1'S)d|?)d$C-TLetֺv}#R3B -֦Th䆒F 9;̇mL'<|Pp=#C(`^mQJ'T+͍r1 p#3HBPh@Jh9G~k87=7}\:ŢG 1(l;HQ(䮀.)`Cz2ʦ֠2ZiiZE,bQVY`AjzND~ڶt3BZn:L<^'a1w65e$Km-TJB;DVE;#d$t)MCNyB*rVi4S}a$R-fwU?J"[tTFp2dq".'\D#TҒe>Es{*girfwlBGj y?ϪEo'DCX7-4½'ūsykˮUbE׎`r;ZWdAa<ծ 8h: K% "!Z$D MD5hWDC(=uC(551B14 YQ$)wp:|wof8izeqzYLD*bX:ޙ޵|8aX+Z}}wR~Bp'{]Z5q^npȲ#÷bW  2H@LˤA\dʂgsP-$5wwoqtX$M>_宝H2R1ǣ9k zz;U=G{ש޵yq<:Wߑ-Cz8l^bG`5v}vi쪢IpeҊOKFhٕL7h-PHhQ;3Ar{<ҴԖR0O"ԁqbzBl2*r"*Ci-X{mqVXPh<ŋ"ˠef~B#  Єy[b%(pL߼+YsAp=PhܵnPXI<$ 34DHS"YPh_1'ߏ?4pmAAJ 1R 2Nbƾ:n 8'{~\ZĪX@u!!'FA d٢tNI`1*ّߨ(>jhB!*gF連՞kMmX׸tkbr=n.|G֫un?6Wl%j;.n}>:V-h Cz '=w4wӇw&wjI]2SVaARVuG;h<-xCt=HF#RZݷ( '1rsy;]tO />R]}:x98O6k] fY]! Mm:Yp4 }{O9[;*bPRe Nem==@ݞރ5`Pc}vPo -ZtCma;c;y9 z{|%KxSPݳa X)\<{svUQ^`S@~hwt3@6$3ֵH7Lw|^sZRUq2xV~Y}<|2v6CH42 ,-V:eػF$WbKGD p0`cڞymZ-b7HQ%)UHGY_DFFD!vFՍ9hk=Ώ.;"ڲn< 2a8e+iinE{DZfiDH؊ɮ5IT%|ừ?ۿ9E^}l/ o.HwtthK7Pc -;kB{;OW4[x{FDg|,XРOސWEN[Lx ӯCIOwFL`m[YRƌ*4rVFHY[@d-R(`2LZ6VZugMH416K1h#92 C&(U"kDK<F ԂsN|0]0%-BR%C Id#1] 4`AXHp:d& 5AJhBN![A ړRtf ĎG AxAߘ6w(%jHs=B}JAT&j˾0ğLF%q=)SԯHf >Q,i`w td)"eYrFg|NAۃ0^Ng3ŗtSt1M6\Nr EGC : H/% 3)A*& H@gGA#9HMuj@#$0y0Y"nD!+j&sv ,-j5]8UrW7~?j*> @gST"mfp]@< BRkY)xFqe Jx=T*ҮĨшIS 2tN@N,U&'1 L*Em:NCۋQ/ޙ۷ek\f`t>hoJE$'WYvyr(z6QFҺ&6iC&G_b&權/:J"u1Ɔ@EQw37< cmvzs%NQ'誖^VY]maЩ تȇ.UʺS%3 p [?cngO^MV~㑛^KhwٟUwS]:nHA񜒓!,pk)|?|7Q/ &naW,䬑sb )]q%ކH}7?}rKHB,NXy#}Wd3JxIi=,s戧|>Ypļ7u"k?|&)w)N ]s5g^Lpf^Ɩ]Hbr5 TR'36&L )6t3&skMT=nx)9ٰes%V?]~|LsK0̙$EY<<=r~SҺOK|UYXLiVzdh+Sl{dE2Q1q26ւ9b;&~ ͸ Ќ *h)b4Q[G?o^ҎH6UP;=\h6{7a;f|}oο<*%5N[t-i3+/H9? 5|6tsц5bdvleJq/5h6J mVkc㑮;'tuwΥ`A4mbGLK諼$πlsE+fQ^~9;LWCQldY" T mq5ِy͛7Y],\|.oU\Zݫgې.>qߛϟ3݈L_ڲmO29]ĵ@:d"5'zBEx]/ R|B1hmP/[~.碊IwgObjgjŔ"լZ Xq'dKYK: dfUlӅ}{e4 Y$I8r9˩!d]JoyY|-}OʒC;tϼ\kVRKCӕIT$Hj|l*x.ZyBOυx:ҕkU5+|~rы[HqK^I>-IӋn!P1! {bq8;ptq_ OGtc:Q5 !aY;Wyb hBGS@bQjY;@V/ybkH+P4rs>`qGwNtB%wDQ}YCv٥Tu٭jgbqT\wB1D~^}@Eőh9~&RG-gT0nv@Q&HooQ v5)n@"JQ8+~,Ww*{<:њBuK9RE#wbmemm +Eihw-v#z[wH'6Onݰ9's7ᅆz~bYÔ򪻍]<)lj[c&2d;kBwk; F®풓vη$qY[xIwON j=Ww&/iqSKt#acHDhƴlf ^Gg=G)U;в)=>k Zqm"鵲)v{/Za[QSVa{kXĤ_׹ʑ.aQ/E0^ڌU`-9f[Rd 2E etZˤ GD&69Srjx&h#buN ӿrmъ5>T s$u${/ڱ*5% [RR=0`fVݱ 1\ U 4J%V {\WČe T_`^nYɽHRpJϫfC՛Kkg`0WjZ#J _nz_VLi[R.wvu,N"R@evRaJmQ^e.JbVBJK-cKb}[[qf O1{+ ر8#vxI8}XPW-/}cu#DDvni=t>cImz[w K~˞Ȏ'N.c LPzR`|񌯉qadm$MJ"O3 rC }0s \f8~EDJABb(R&#~~(_26XNFs1^FBŶ!u7T}m{^.-sPƴH0ӄF8`r2E#aRY8T*\FeypƋ t/V@et7ϓͷUQזy뿂EPns.ʍ/&;*k{v7$ʱk(8MPń#n E02Bp.zc:)ܽ;1OV(Y d}&L_ O;+Ž\ 7\"Ez7CaK ~' iH!H ȵmuHʓ8O(>, VtT<%jkYЃQEC-؝rԝ4oV&ҎX}\9X\*&xTBTzoa}qG(a&;AI?~X4fϾEKlF!gsq0@Z[rx.s6[҄6z,)>UV0Z{u["phjm)\bj)ml(^ B7m}2[.V532<-o4u/K \曆lb` T*amCtg a%mDUktOrx$66fINC!ء>:kEg@C 0D$KkUŦJ(#@{ǯܖy1O%$\j3 SZMm44is0 ~ҚNYTgq']<ϧ(]AEL<5)^N'Ot'L"#gd1T;l$S(ʃRub62REhh,!8FD`BF )+Ỏ 1ְ/{@PT3>$SS|珇VT$9.>|PW!y6ѕڇeWCWٍciEM_!x&j>&AbA2M6xv8|[Jbv2͖ůҧiGI1o{"m{I:\L'ɤ Wle>|Ho9bBhG2!h#lJf Su@M_ڒ8ʃvrw8S*(,̾Fo^n ./ǽt9C)q3eOvW0@BM cBo\ZYҖ CG0黠с3Â@_VXChcՙXTpniGӢZ{KAOSYO>ԵnCkjJ!dc}S!_sN9̊UDEǤ !2@O& _W(Qrpl,mSFc_zYhN3O0X*KV;DD2f˦Vk&gv(x׺t B(QY{Omk qWMSD)'WUĺSviCL?H$1m=Q?X_Tى 1{Q4M`-)п3 &(, )%ȋRp6iBkKPmf-Eq✚+ VLpKI9IQYLRl!c$ ddm@t&zچF:#.1E\W>ҚmSɑ;6հ@d3宨M4,PgDHQץa \b%1e>}3+{y~Q#Џ4F. ^vƨmi _8rsݰUAxC:PD}c_4XPE.S} G$*i:԰ zqc=ׁB1[u᳎Jg0TB I@  NF%N`g^v'"ŪÃݸFݮ;JbpcwJ{f3{f ]FDDNi0)[`ߥvc9k`7q 9S%S\֚!;v׊؇k]ՇSSq!$?(ufP^k+޼Ns!H@5_X>No~X?| TrR 4ϳx2*NѷYHn"@>%{[TCZ2$l[eWNΥ`YwЈ6qU\TA83f hn3]gE*1 _pmԼQya%џ•11E2BW4Cp)2On^*'@TP^3[e012R.UMB@Lmb$MyÚD@@QOs* k !EhF7sib_;$Ϥ3P Ѧ,!2V*`Yw٪YǝVהXb`ѧPJ`,/8Z̥МfoxOQRAy kRTBGY*V`I$*K8 =-a2[.VOOwHk!T.{}! zLnq4A\(flk k(ƌ_3\5fϧ@QD=5!qt֓{NC=;`.=_qCRo3)R^Κ/~PY1XiAw:-ٺ"zN~yr 8RUt靍:GQޅl^L99K8S enGEijZjyU*҆NxgBԬkĹפּԨ+ZgۮȱOJB( b|b}:|16M PLz~MX$K4*`YXy#L_\בHR s$H10_ 7۾-s.(y55dWsG{zjI+N}k*ӊ`LĀ\XD+Ǭ[-FgB$Mv֔/`Z̚j ;ѹTPNEB͟`3-0; KW&EHiW1z$ҪxD.% $ /% 55'XM=/5ᨥsRA&cS7UZfZxځYgX~}jXvJP%!]7fjN&!ARg9L8e(05?LߗlфK8ϫ*L:tvznc(S98]MaRQ.=bM}:D$t.vM5vɷHHF8(_suZL6(2 k\#d;PV9إfL ,|!Ȥ92v l G9I\/ Ҩ2ɃOO5.!C-;dvZ<mag 4D JH0aa vFm[NvB bz@2;nKAb.s~&{,7`gw7-rRR5.ZɆ5.S"G0RN<] }3j6ilǶ-Q0s\Q,mž欍k@QhXZ$}mZel_+|zIvuU0y餭|,5/!qD(s>|19KrhUyVd͏y8GۣPkn{a%UV8wi1Nӡ"G~8Q {c:k?DlcTr}>q.<1@c+(H|k)3&vA!zcRA?& / cF$t9s$}8\'_=ILlc)+zSDzyz{SZf }y[˄;P{bQ_X#Z.@0E' n-2m]oL-5OcX7*ѷ[&׿oeUX<2jgcV)x,&rf:es2J`h~gͿ)oi7<{Y̳[>"g gF 9]6}R>}HL˟?X/Bjl^jAڕ.~ږ_N*c,ٻq,WT54TCoڗݭMSEX; %KDJ@@2~eqx~$*~ C&J315@N&+PƂMr/0#Bϒtd?qB!5"SЇsAO_%;Cm1.Px *ÉơzK(q@f W9[[ +Y%)Vh$zH _ki5R~(S`(H yԨ0ܳM)b1[&S[!} L֭Y7T?l8l56]t^xGjA%}p*] H~Xݷh"iuhtPm{bIkqNJvuaEttz@χp;P=c=7@Gyw1@Dh(dIߡd$fѫ&t/^W.$ ((Jc|F2%X *ӴЩsEeX/zU _+^ L/ln1.q;Hd!6Ž*pD|7hۇ ~'))GW|[jMǬs |\UjdcZUݳޖyus[]%fg~?VzQ٫1blr.?uNn6GӲ4ew?֋.R^m8;6*Sǚs^8Qw.EL_VRa[%ʞ).F<n+wN5Rs=D^MnwE7黢]MMڽ(4Iu,͌*QDTebb2 D"ifݽȕS"WzA)G~oH-䆔1)''`PAQ؊ɯma7G4{V.!ZPk#kmߧoaR3sP8ŖIJN:ƌw=e]?fV0Me6#JEq dh x(lJOW(k(]ͷ#O!|z>(!SچJ#!=jFtpY|Am?1չkчGV~nQXX@3P PdVLLℰ+\i<" c^&W[+{bP]*P5̕_zP.wVƘR? Nʩ1J0ZOb;zY;S1Dr;{yNZyXB5%āJLhШ)h߄hAGÀN_GUugپY9W][lߪ-S{<4sͩ\\qbP_}Vc.B~ ښzVJJTR _l)AmYg1UvG÷|¦,ܼy *4Y5)z[->Ξo?[4rCJ7k^iRqe;SnhWNOC"@0_q\ðǷ kf}lwkD.G!o;n4}-r/lRh GwC~W G!s7<ژGK3(3XT=ca"1FbM0RR2j 8XÆYi{tUa H*ac NyV66WaZMעU],ǺEљE@6,|U;<@C5+!DlPba (Ty2XF39›́"$B6`.S8}<<#hj/XAF~uWσZZ3mr*^~mt]Mzˬ,foٍ|ftur^,-6h\]Blyb!fP+"k5 Fi(I&QF4_ڪkwv|xVPiI;N 62iY ஺p bIMwsi0.x_q=Ux=ʾ!oB(G<3UvT؏Ug#NCCNx%^ BC%OGFt}| ϧ|Te ؕ?ذyMb[ݮĞ:am$MSoh,^S>3{55&{OeW&ּԵL_t{0AhOpҫJvX p!{m5sO :X6Lȼje<.,7Vػ8|/uQ H{ SE⵺ 8?o# a q\&l:T]xV_7PeC8A_PI$[;) k a_}0|=C<.7T`uZUUPnu6( <8Dx M"a ؆" `64CōP*P^퇑C-/#6 gM0TʡJ-ֲkkhzCE%knM̀& 70@N^+u~?ft#7$D) REb^`FcM3aLRO+OB"sϪhz-\GBk0aJ76M+Q9iqOG\ )\Tp (:yDknHk˂n}. #u4irZšv[I r=eV%g;A\&~J(R&iIv4Hișei ,V|tuUWWԲW}冞]gޗ0KmMZՆgt@ g^朝Y):GkXum^ Ta3 TvϬx-9ÝyWP wemZ(+)^t .v$`}\؃,٠,h 41XyW)]Zi~ ،pLZp]wF${n9!1'usS?I&]ZvggT7Vzo]2tY/;LN˅kW֪2+p2}I;V&2=aª:jA;K8;eUl^=;;v0khAㆷjOnܒusiN wW] גZ#窂矌⼴S_A1|xo@J*fܞ[GydFbFtھ"λNm*^?3M>|6M]Ygtڋ$ߧaЧm-Y[V_ZkAJre>e6-2)fvdJ^oDžMX1,xY.GR+$vAmz2mB\;2ZiOO#Qm{{}w<8 `i,u}K'Ӝ ڶ#w޸VJ!q&?JQŬ4{[zXraWlXJp:o֑9J'"d#[ !Ѽ"3L4 H<JrK.]ԒT'a1_EQ8+<_0 GT"PuU ?8;= .o×ZWňm"?E;}?_O*k"̓ZuV})mqo77+|VJy㟋E;E/ưcDTY1`Xl^LRU_ :kHW?]3T11&n/5hמ;돳"YPăb</nw9awpPe Y>!zrHbٝwU1zW3 KB5P*}/&Sl<*G,2+Ϩ>;)Åi5RBha:7n2Ϧ H}i@fN?'qt 6 `#Ay#L+0IIBzJxcq'՛s1΢`9,.&_U^ER ^16iA 1&sa$U֮dG،oqVdǟ(YWknxKߠqHMq度bA@;(\ZU#}̡>c:cyXGGms-?]L j&Go ߗw#cIGcjj FھGKfɄ2C7BcIR<Tk nt޹5r*5[>H4QR\ڹ?.(_~nIlF0Û}Fq^:>ɁGM>Y'c'ӽODt:Ĕ<@pIIx |8%;ep.$iL2G--ҰD0+z[|Y1vPHӼ|fy|>r8?.w Bh>'~ t'Hd0r=P}W>JPb} vсT +d²e~Ӈ, $/xΌwT|kSrRwɰb%EG~0W`*k}3síNSkՇ[ w OTG컺R62nL}#U )Yg8L#Nd '#o+s~P@u`Bd`#XhFRgzo󷕻SӈòK5ٞh>i2Afߔ[.tٞ `٪,o0l(rLfagos V ZȖ¤=~LZS1z`&6>.0&Q-^׿T`B2kRHcjU>iX }SfTCY*A1:OϪuS,\4't2(p2οNr?3)p? x>?=8ȩ&v|8o%,ppA0;>j! 841^P}67y7=-z>ߌ`R4E!np^o#Ps_&>ZʩPʏ>;iF63A3"x=:XzVE*MxʂQ0=:أ=:]*tŜ*)F=8؃h>5WcntT~| |柌|QV) ױq29/~x~P,ܞ8bsmmߚp%WhҘ~}fܿeazzT혯 K)EWq\tAg\pM F2#T_6sEr4#o>ƣcل0F*,3Հ,<]QOd} 8:Q' 6:Yn3I4#rxH <a/2dUQbQaAk[$LDK<!EFJbyC<.FhZ "8INc(A`Pښ:k=I'AjĒQ{8ĺ*b+Y.bfkiBP87R'Wl8+kU5/2g0Ft= _4 0F&'ViQI0LWQlmGRU>`6Ps՞}zc\!k7^j ]NɪUhUDNeK^pTHA-fDznpu e $Dk@5%l^7w^33>N +x2sAչ;k}8gNVœ "ӫMpqÐNVI$+xfԥ.ؔ_jDs*;O8xy x@%ExL|j@%%NUj21$" YUau Zru6(k >Y ƆxjU*L*ʼl\'X󘇠}UH%ռ%Um\C<i^5 JN7ߝ%'ݚ4n"}#Ed Uc2#rNۃt`18 .>:Oǟ|V88?Y*W[.4ӣfko罽;Yeǧk^J֢B<P~#HDL5o赻dw~yWg۲(L/YOLUg۷ow]ޕSCwv!:lôq&=Y#n=K@!NVѕ1xxxQG]L:R54 hۺ-Ř-@NokQ6OfvKM] q'@~{8xE%){\jZ P-*+Qϕtkj}^򈺽G(/k{?hk(%*v\!nFء$/x6uDKy\E)"8n#q2KE60Dr1x@L@c%ur#U@6mhr;)s$\5.e%MۨdWWPBTT  rZQ@㇊P,Y)wg~pvL.U7&r&"\ 2*O W;1ZH 4> ¿ B]~ Ƴ0 wlb>-\Y 4~앢.N+Wa+볬  j.s*Z,Y{4fsMf%#;b~YK#gw4-r_GmU*Jn9^^)tWH oP56PcyOZ?FO=v"0G+0܋`W#8 XM+w@h.ܨnЁ{ŷR\Q˹JBfx׋Ž&%5#\.2'Ǽ;&?x-җUKtFAN뚈J=qrdwƩsHj)= :[{8 4~tOPqyvD!Ea!:z0r=Q Q3N"g]@##RiYT:6/+h!NZD' a;6ͻUbUi1^kZz\x8z1(eܫnm ˷#'ߖX @U-56+wgd3v&7SC{lpR{*ǮX(@Bz34B#+cWJK5z LdhLhm4I9AB]dJ2]\[dWS*DA xt " W&*oΉ 뭐2]JW-mf,HjnJt -]lD*! A6D8K 7|ƍ˫v$46 kUuD$nTn_dz\^BxѳoAhjÙ":ϿxQl/9'bz1;Oڭ1h(U?,CsF Q:aTEx4S\sKRqڥ,^I ϸLz&Rd1 ZNj1}]XZݯ/ȦgV4iC1p[,- c|98kC &}3܄Xxtx˲*s٫3)dYU +!m`j#xt9{\Q9t!!'T 6\NhmE' 0=lt=SR*I2YW Cb@p["eq* @f% f6SQp9ZjSks06섗}4_Ur6j@ˊ*ⶣ6(*:,Q9J0Կ>S4_;EXg~ZIwMiӉzFLl4q&Db3ȋpU%9.|'M,VXЭ@ E|xgT!@ǥI  Tz8IX.o\nJQX:E(J !W4"Z})djK΄Yfj²NGW۷kVGfc&+`cK | A @@Q0OGHs(7rG t4yjGTIGF W1 }ͻmvfQY Y+6 km^RWog`n{oj,,ZoQNύڂs^{@= pI=iD_MZL\q ?4ʒ9bt6fm܆Z!x}髿+"tl7vE.rFz 00KxL9%2FCa-E!ĈGӐp3[X.=/A~b"d߯Ofnʳ8 nzR嗓羙=qU4Wl荈o1h<`6- ǡ05MqQ%-a,ȔPTeI%㤒] 71~4L-3?SSiڥŒ]c>g6K>:ÈgŎ[hRg'kSee`TxT PBzL-Zgߦ^]W# z]@j5b`HVaβmR< ~3z1%|*o%dj[p8k"s@6:v5h KQ̙R!hxDF[1JsA(9iR>mCh%O7^(/y+Oڷ2QZGQ+3{b=Q`UnA/zd-M$\ Uo 2t5X! jۡM{ H2|~U~(W40D父fC, ٢ij{(ڒkk~Um?{-M]K0J>Qw\ m`W,h=QAj&IaVz2jV2w-`RULTjK|h)j a xr}Ic;^eP (*p0uԂy=V]kq$"&lYu A03,EPDddԤRݤdl}֫BSs%" 7#WdNf΁Đg/4"7R+NܸpY2K!\~5a;-7oz#a{_chC ! J%L v̯Me;6mY-y/CT6ր!]9U$n`4Y8O(-t*=Fz2zύe6D+<0"|9W;Hq (5j]vڕ܄^[oýU.v{#]ې(fz#@)q1)(^mP U(j"Tb8sO+5Vº~α2($[7YV:ut-9e}kرZP{j 1| IDַ= xR:1(8;(ie,M$.+}iLKm߼v ^[/w,.8tc0Q7]hW;JY?[A&WԨ<cȤ&$@I ֣ Ό ]e-\g17{2Wڜqݺ$Ҍy$`f,4ӭ^qYc[-p&!Vz7! 9Q(Hp nj foƵPھ&VRa^-+C1;S(lQ-B2;K㌲/ؼ.v>~"/YEmhA/^̚ IZV<ȩ>Lq~73~퍦jLb*ETpaɖ5yTn$_\~C4YhOƫ؆:t~?Y~7.ɬ>b9/ME#o)ֺL&чj/U՞jd[HB_ա()Z(\yW.[ }TJ ۵~T +}GjDҶ1k>c9Q3&D+Rc}M&l~ >MSa)vd7lW}V(2:o48YDg m>=nUmPAuaGF%#9u.^Wj \) H*˱JO{'5wm?-g'C: 8i n?X[Myztc#? heUwZxl-dJ'#JT6d%6>t  $\Fޗq0G?VK}Vq WC2VA[^yVXsϪ>`%̼L8 +ݱ ;Tϊ(yDP)r0O74*rXɔ5T&,O/C~6FΦT\Q>R^od^KHF)^1,qh@8!SlO N_i99!XP3|{ ʃO촘 izZ1L~$?0_R\nޕ"<˛?{JVd4niG ѢmȦ~Dr@ej\ղ}³Ox7+׭5)tP\Mn"W["etlڒP] vצwE* Ơ<;,pV=:t eȞ8_J! gZ3ܛ̱<{(Ub4~+,e2U暫>ƙ8#pQ[&xie0JQL(kdM(Rn2IqZc+ xI&#MW\Y2Uuq.>U,X-=;T*B2 NjefȲܼsnr ̣-b4'?\ߎ_g$VN+xlgaPPd6g׷;7,ե9s݉$"m RV zm|@#,qf"3*&&?@*8pΎBոH"D_ٹq*l?vCa1BiׅS~0^j.5 ) ^s-uRJ:m;$ЕHn+Ńp6Y䵅V g=f^,*^A*Yx \q^Nȣj.9 uߚUo{1̅Ί +ɸ0iJ 4}XIP#IK58RinC̃5yCWbg^ Jw(v5 jӕ\=#qJ>|0y&Ia9O8*zd5FƻWه'B%併̀wv&ĚtA5jײ?~?n>]}iuNRUnmjn6\T#V8%ӷs}G\7T_%ګHFT%1peγ}?8SxĢ\rN9h\+[ro姟$}Bz 2F*I&`KQbԞI $>GɒqoL6br퓿.l>>,XQh,[oeJ:f\Dh;Q[qVp%;p:./$J@\ȂH r MpƱIx#TjMeڧ4 fO%ִefn^&CNVrF2.Et)G.N4H0`@h!*G|؀ll4 Κ0Ve(b= c0L /5tֈR+&mQGZ9S腢=R>j[%h79guP K8&[v6JnA/-9Lޟ=+3};I%xs;;SOO}Ǚ~G1PCQffdnN6/[>QIiy$@3Y+dV&U=1~p>\e {HfH 6֯i`II?Ρ6֝~n&KBtvyϴ!CM%zh)2h\}4I$C! C̩`Y9*,wc m:G2jo}2srY{7rz]MsfcX j6Ř(AbVitL^﬎ q}M)U̝Ekgp6)t!$μQ hʍ<-aQ51/mb7ZkANt1_̲Hn4E%/t$YbVqg!\Fi"7^D.\̩ZLӲNǍRwǬ}(YŘI<79b# 5aDDSJ!3>DQtЁ)*[KC 2 CZ=}5˅jߦ:7B< i;X?/7fdH7M?Se)1%i ":$gJ ( D2cƂ@A-k<5D`4{2+ګɲϞ;zx]wJFu5 7卟my)QP0weÃ^5 ܜ/e;jJbf^zɴ{)ҁ+/T "_g,8G)3c4IWP.u0zQqi RHͺCN״c3!THhth3\SļBI4bKHr#Or Sz0?^&g2N9uL, ?nMg"V1yh8r A,rqGSMPʢs2q:g#H*"a" 1 h@MD cgh~^|Їx$EZ`6cS 6*<1V"AjYV'3<*Cw-46RфqIW>ZPpg leZQHa!){8\׶(n[R -L;ś;SnY ݅Fnx!;oo3Ebv~(hRI$;hX'>P:C>(djQDUujٹD[nA%ognFhQ^Z5?˽9 rMIdnbmaEGr ElBՋ[Fg Z;Bqn[eNp+*lNЪ0:MXjʩJA!!:,8-}2g2 \DT91Vaf3>EG+xf_J̽d[r&wmm#KyهTWuŞ9؇`ga0BbYql$'߷ZlZJ<pۢIvuuU}uUZO1glqwCۊǓx8U#A#rM׽cك`[1]7oN*pΡV}0X ahZʺmR"r 7_b<-NMͶ4"%b4ъ͠ ndw~FyW-x]{ߟW&oyƣ-Ƕ0#F-hkLC[:DX} E#sрwS*SO@CI0'ڐB7bk3ebL;$cG)ށ[q =7.^밎F2tכb=>0+ ]F Y4OOe -"U? $m6Lr2"[6ƛ6Ƶqͮ5t߉O b]zRrϳA̱DwSv~lQks1┬jjw 8. :D%8cd=zzϤnO'xD#qo6o.[j-So/\ꗸxv^cybagY'oJ[]/ ^sӺ& ?U>ùX)k[BhW˼}Ħ|WWof}Ӓe> nr 4+JS_/o"=TlX=k>;RqDdްiSzq/)dkehJ{lb;D*#ymaVBE@x0X T9yO)59F,c3=Pj*6oVȵ_nsS5cS显ϐt@y5i]L\5eC[f͙%[ƴ(uOB/r 7j~46"ZxIɊyb[s=DkzaG!)}pim!7RY[֨&|_rPLi!Y,%k/6!Y<9DboAȋa5iwuɅ}a%7)+ҦUz:doG7ԧp|V]Jy7GQ-95'M*K.*ZenS'`CWD#XO!_U抾o(F.Rdv%9,J֟츧6YןDؘɕH\@{32/TܦԖiSo@t)}"۔fyY<^lbwD}%;U!Jkh4o~~ӗ$.>,;߲/*"䘚x8V{#NxqE};HΨ]]o? hF$y>'ui'2> 캧ֺ'mUT> ;MRO7r>~gX~e\ђoI3<ZjPRV\̾0>uִ_M=6z,EK,L)C9^'C,f2U<2 .{Zךtr5lL)ۃH0"([ݣQ!qڇИ1uAcd|q4 ,Pb#+C3,gT>B]:2*UADSEؐ6>>4 , hSFR&;PU0)ըYnBA%OkCՙ2p<^ B-n@h차i_km؊w"IU^MXt")ٰXʨP'r($#94r_ڄM4"AMr1"cEDBddkTuZTr~5i&H}s]%Zc"y[Z%ٝv{ͳk9 [vxc=VVtGɎ{ZjPs NaGPj*wr"v6 ln6EyuD[xkP4$pr.W7P<]Mx}U%7+DZPb 1u+G|ABeq'o2Ëw5φxoؖ xkI.ނYbݪ)=ڞ .ZI:#HBޛ-N͖7>LL>5 Ȟy&FT|+N֠'c>Ds 9ҴkvBM)xTEBarcr!E VbP *Pʕo8R)oGnp^BO;ěW(&[|kڝS+|{NL9z{ͳqR TJC]XB(yOfGJ@=8I1++Ldt䔉<]ϳjܠ/+* 09rξP%&zKdz]P!gx!sZ9oq }-9׹8Z%m3Sՠ"a1L+>5?3:k[ Tњ7-Yi0K嶧|X 5=Jd{6^j*AGҦh%OšO֓iwTӮDn29 2Efm2:ڇiIT`w.l>V󼁿YlH* \hc̈x9h1QՔ." Vxt,MI|E_ۀAF^AG̖WݠFX!"9RfP-^^g^Xhs{Vg;:n&4h--*f1O08?x1X^y>҅Х_oei--!_QƧ-NKYLh{c\zه(6EhʻoYD&%9 O1arDSQ{muYn=*mch_puz#g^5N)((^1j|u1 I|1 a?Dk!]`˫].y61X(-H%Bl-QJFcU.ri^A@՜\, }ORn2#W ZGXy~Fݺy<~,uy,*L&OU[NBGҠhYS5y:Rc.)V9zjqxa>~k5<26찡TʹubcE*EQSJ`gzdg6+WuJ hgL!4NS7kC{QdG07DK54ELGҨh4=K+ppW EDyH(lG8x>mq* lJ7Ty%T֑Ru#J Y#GC6`bdF#Ɔ!!%߇;֘Q&6U ӦIA$EXse@aZh(_0qCFύN,gĴaeR0P%B#Ugpy>_>+ٜDZ!h wڅ<@Sm^ oetZulw-q:vxng19ݜvIęenJrVU*U } w $~3u'hĚfǷf74Ydउzط=p6۔Cp*ki! њUw4;iă% ^m_ɱ-GJ[=nT!]6g2XcѾ8 o:&ۖ/5MYR9a!5ia)M",Ԁ3|F$>j :Zyn(pn}-[jQT#)H^[W\ҭy+pRɖmf33+Qqf-$M2u/ï·&ϯ~Fz} Y< "񫶩fۉֵ7չ7.KIkl :jFD]ή.$TAd^FD3FQӈ ,1xrRԄ7[T3:{0o "`xRsPS:xx` HtP{c5n/|ϯA!J93&6A[aYC1?=FF=)&#Ĩi/7t ˑBj2Tj6VSooW8.mvsGZyD~ a ('ɘ |IHeHwWJ{6FF5GoJkjZ yam[ǘ0cFX|:Org(e$2ׄgZ\`HLD;}wXE4A9KZu0:ڹP—+α #/{F]qIT7ZQ s`W=p*fNcr \f-grRAP=b/Pɨ 6%i4VlyOY 1'7^l?odxRj4RLWA[tTb k17(NA*{dyV (}fY#OBzrq<-&JfSDAٱBe3G*3/-X$cn9/aHb8ĬLh\!Tlw%\_3ttϝA$^.$)S.vB|/۩T9]۩Q3KUQ&nU.P֑+'6: qi\S8l#gҌ&]]VԈx}s庁;pLssۿsf2LdbqpޤAx4_i/tiT!e™;>(s[Ӕ4={fwiͽ.?. mzV=\A).591$= fʫdȞƐiT (h~4 DX)] ]K6U>dZb,<5Z1M-v"CN eG Hu_:;P{V.zC|ڨWBEm91{GAQmgRbb؎x6*hfV):[ƨHg ǒ#qF6Rw OWR~#}WHg6Hw30k#=%H .7dNu@ʁ?ifs&XyCH1}) BS+~Ҹ;8{a?n0K \oRR87dߘ7h b9+"K6O17}c 0:=i\7lk6@gAJ~!π+'lxtãc7Gj 0)NIpJݦ/!Ÿ!v[}3&Pj`;RB)%[vsvqąc釈9nB;>͜`^B>5Ԩ|\"M&/).A쇔6@<үaiNFyNZyn*6@z8zƹ߼3xK"jRwß쏝)1p#ˉ!C04,RzTJ%bnԒ2jHՏ2h.Fw#聿_"er%%I9\"  !Ba+,49l6RzwDRl1Qsgg՞Hh4pBd`@18xwa*AY.Z389ǓΨxvusNu76t?N9q;S6$/RCmMBo+i~a!C鐫ʐ S1`>,K/CLEUBʽ!r !&SZEz5#h55(c)&H* -đ\}j bmkc9hJNZF#kiR,QfHMI[H:-G\0(vu^P6˕vĵq-7B:!٘?7.3}Oڅց+֡%l,1z«ճ;!?wQꡤJ3jlc4(h+*gN^e/z'ab8՞}rR2rT?$8~=T::Nrٳ\ѱkKdloCRIrដ|\cvp^Ui'eNn׫r;ԖQSJF{J.xqQѤ!tCQ*LXQL|j yH |*&!¡&:7^)D4gZzHr_1`_v=,}!~ɲ۲ǞPR*\lKN )dG2ȋߚ+SQ_Җh{4&ZinҤlL4+;*P{(UЅ*6 vJpfJcJ%r4hhv&( 6gd]P31zc/ogst3=& ưu<݈q¬-if@qK{uif\t?R]FglO0k{K;dōKY'ƭH5Ox 9J(tz-əklvweϑ(-9]6wWLQ?Is-!)qPjw\QUhc|]uڱp(d0TqD!gF1;#eÅ5?ysk~s{iG D#IXCk~hξq;%sޣ;t~џv3\!sc Jř8pkreDɧ:*w)r&VKѵRR|-$KjD^~m~ҲhIz6pש'n޻4DCf~JMoΒo.ڻ$+d1s/u?N9ESK'Ô&B0KFMJK) I`-n@IJ5p*:7q1 Ut$pt b@fFC9̥̱,9^0N%Vm2W{fG䕹T8%n$E ѾkU6{T5FL'D̴œbp}5ۀ%>Xe4i^8βKV K M@ Q}G ]4f{/Pռ4ƒ/o&I!]#lD1q]ԯty+0e1de>Ǽ]Ye2'diStQ3+ ' maTS=WFZ!rE[cDԢiFh lH'ۄ[kZx#NURb%-9Og-Rk $ag΅p4b4J0}/ÁQu6"eFj.C2Wlt-p.&a yٳx[…޼u7ǧBThI%X4gN%DV[8Nh5xDQjZ&$R튩f1dP;;Pw+03Evf˶,#Ԫشg{dm%TƩLCHtg~6QLdg2Ѩvl7 ɷs "l(SGzB˭n a_gcߢPIƉroyY0=HS*U =$:O9c{ZrGR3x0ጮ`/Tu.c 0I8;e_&|:=jbed͵Շڽ@9[00aT%L"Yj\@hRˡFk5:>aOX}G"6p/Xi>/Ć5yѭ|\oA@솛FKJP$.y̞*aa\UG=<$! X(6x.jtlTqrQ< [AУ{+#q\d;'fk2zt>iqr] >ĵSc*d7̽!g'$T<"1| ̸V#x#=qS9t%CkBV]ذ$k$=;8xnEZZ4|79l~Qr浏~Ooc9or$KƃP8tzͣI0E'ff؍H(zj0F8Xsf=B6ΒBu,*fSIQgƩ<9w.P{ qk/SRU|8yLXkZt{ʔ̂p^?\RgD7}S{e ^Dy J8FΠ?glS$w[8p!Պ^vBso&7hq>?3bcLЎ: P,+Fۊ]ᵣ+u)eC:v䑨P2=27 R8Y>qpW 4p&ϸ( k$ӄ).JL3xǛ>0>{in<*:F3+y@Our=¨Nu٫E0+ T5FPz<7$9sMB@ȍ.HXS&rA2}Zxi40WтhelēZ3׼4s'NPakxnϷGQ#mX#kv־ۨ ZC3a7`{bf1:c8k8`|0I9#Ma\oӓ˝~l/tlJgg?Ňoo7q~'H_ HslFp?ɫ}ҊDW$:kߢ5Бhp도D՞YL n=l"C/P LƷd&ӟ/wZ9ۂjO*}^].~ 3?i_G# Y=^vvg ybLeZxWU'ffұm"^߾_RD\m'nM RRI]} _yZv6J-=!!ق Ѥ.B4FahG{x^YE{b6hGLou_*"UķoXG g:6&䮀դ"?;b S7s=u3[=tNF'GXg'vڼdWO}'nL_ҵTcyJ ^NZc+oʰjPu jWU&E*J1nةȸLlGr4u @6rE.o~ү\(4>jƟ,}|?-*<9~sNø}R}bxNkc}l{4h!{\0&;75ͮ:1fFX=z/&Ʉ%ЩO<-0]4ٷ%o UCxBmãEL1hjOHBWВY& w)qY"TAPg&ZV +238B/#p,Npr&fD5NDE43 Bc":!@#,4,V@*>4ۙ|<0ش<-f9]׃pݽցQK6uH`Zз~{"Bua&Qtxx\`@jyfWp")BSv滶h#")DhLYvpw,;nl5n/>i[Y"Z# A֏cC +arU%')|IJ! !9D"t_#A R9X@,3L),w'd'H"EO:߈8u&D߇F$ڰƋ$כ J]1L)Lx?IF[hb)\'^t",'kt:qsG'Ks_HO-~@KEb/=ZfgW &Ԥ*RW,NS|`<,54{ΟJo OT'^h1n\ a IqxW2Io^2)MSw98R/ rb;6yƤ9 i!9jlKnAR7g,4p Fڰ_\ÓH0wdBr| OR&`!D)ޕ]urN Y"z%ЩWzH*45{Fhz*^># Ӊџ2Y-'~ ;VJ)-r^ jl-N$&)E^1ʰ|ϻ?~&à-M?RWC ڞHRT"/4?S!p ҉lWd-%0 v F+ %Hɓx5"N,%˛ Y(#)jEKTW%(T^rNQR^dJ/q(-Wz(N'&sN }qE5լw~3T"ɁUٿMiة[gy3JW|B}x6|u8Q7ȀW|W]HHgZR ) fÁj<ӯ/g'L'ik-":^ ̓2A"DPf( L\ ETÑ /4/-uǍ3rsY>ܖZ=_дU'o@Ug{[X\wyGٳT])弿ux;S;?t:Ul[}.P< qsF$Y6 5Ƃ˨6:>;2WD9a4 [cWcS9RS4Q_7 #ugȬ܎Q-Ie0ɗ3?E s>6^GZ8MYΔQu\Hl&U|z9%gbkRJ\Ŏ{0Zą'3k%L!rL¯<2A]F(.e55AR@K,@[@ E^PB`,Mw<sPKo> @_wb]âΈsjГY-4@+ize\ ̀"CR0+*C11i9˚7LH #mfh2 &,bც֌iV|~tfLRh@ENWŜP+}IUlٝӤ։7GI;u͎rJFhvћJ(¤*0G#դm}4ͰfM3TC8M3wKxN:b\ Q5S8KQ2\U]\hܗGro Xh7;B|KJ){˩V=T-~ybťc,N\|Z<}?o_QٸBB@ Bq11_yo w7yMl}Sn=<:PJrQ,VE4dHwyW y5+.RQۅ ]4C,^ \i۴XkM=zùY |R̊f?Ґf_"‘Tf"8 >L!b@K񐭽Ajݒ6񽴡%B!~gߙQ3QvMk,ӻ`PaSF9ثc= "HJ`jQ }6ylvӴB!X*&S6jڿ_nsZ}jM޿F!a0X_}"g~_wfo׿=<Ҭzbv33( ?EZ&ԖY"ƞs') b 2̀pq-HK )eYo@OCDž\n>3&:Q;~Fd敝x)\Xďd_;`4 :0V^}4o !^t͟ Cg҉ x|&پDq ?27ˋqÿS5bB.&wI%XG}uݭܭ.z} VRvVlOm O=Y]~=@_Ka#c:٭'gmlcVq'Op)9OCM=G7\{Tꄓ}n"X$ qx8Omq`'BoȡD*aZ1`+r$ƈ1DS^)=n:7y?n}?~|czx4ޠ7~e% bϖ-3w3* izi ODuG5tKY| }`JxC{b$[[]P>ƒ}ۿuMufłHk<G#Ŷni7$9c ?sqʶ Cb4kyugpY6f7f1 &&ܪ9aQiKcdB]aMl-P))z"՘UjՙWBbSP0gDYMa4yJQD9E(@95Q"!E%JQ-O )醜^*^ޕ['1(}ߏw `D\kR(+fE `jEo+|,TVLծkZ^j>7wЮQHH?{L&0) J`2(5TZH#wntjZ)'ņY;+y5۝ia0#x7"X~ 6> ]o Fjo`!dn#I6f>L&mdBCyU 8WpYsavP0-A (]S.bBjkR"Nu2䛖rT޾V\b <C8Pi!9ɰS$p,*hF@pdTlQ`pp+*Yi)yE%`$XckQ骭D K\,^JkIunF}?^Af% / vxDf0rz#[{e6w"Fւ\%sC^F2jK@R`,'Iii6'D#={(V{t:u4yZ0wQ&,YQU4ҥ3qIk"J1U(E/:\a8_†?#~й BoЛ3+1YH#9ؑf!0E1k&A"4N@6 nĬ ؕ*j$hU`ToP]U<ne!GHVws' 〥Kǀ. lFF)mG# clFDf\0`6~dEQZKF o}fĞ'ODFF@noxj,^Wիs`Ը#+ }:j)X_ Q!mc͹b_mqi%0%3Ӟad0;g߁}+c) ԖCFZl!tY5iLd(j?ty#mdQҎM^Ѳ]SX}.o4eЧ5h#|$9PDY=Ű-k1lZ 4 Z ?;MH#S67$Б|bʚ5Ct[k{{ab_.n|*W('9Q̡^FɃ'n1fxަ7G,<,$ϛⷂ =^Z׼o^r,QkwOdR\徸zJ+ݜ@]j_M;z_08 4;4k*тӧ݌ғO#Ygw|br5A[L7$:6AIpz&ޫ6́b.eok,὆:o NU=h|"k_:Rb*y۱b.:1Hu#aMߔ94BLQ3B9(چl訔$B|$KD!# 9orWD]VX| SW? &|{/ga~Z%A.aGKH_,~7n>om]= oBo۠|(ԍcӾo6Ξ<dMdSzYc96K]_CE uz&FPL͎JY&l(i ~C^5KVjvi&jqPFnÖC;tp!ϝ5 ɜ@~۴M\Ps%M;yoVWY3Ϳ- ֘>uˑf=ƙC3ZΜ@oN@ff% 4jŁt5VEjKL8Y[vkn߬_Y{(hV[iQ쌻9 Ǵ#ŝ5+{>KMinҊhvUڵ"-aþ<;)YE-QlD.lU4NjAIꩀEjA>*&kEg^[}@xd fa`>nC|CՁ`aWW;ߺ_[ oO3HQǿ>&6=| [Fa_}((7w w3Ov0c r]ąaSWC38Jj&>nMk?7xDe-yGf; b Kcn:jd8ƌcI+J# /b3(hDi/-JcF%|G6*d-*9bt>=#/v,`tg.kmyAМ/Na9X&-UHWDh9gr݂j\VVOco8e)'r31¥gY+}ܠyGgՔ ʅvXȲ"Ây`0#e"-Eەr&HD%=QwPJqs2"ClPWw|ǝ^C}Η/\䣡L:+iq. VrfeyŶqn#PdLCdb&8=mC^W@ӱOzbE6qC`/k R!bرo|.WjHxzyUOu9~6y;+k:>&nз8 9 )(dSߍ7cв7t=x3|ߛnoG~7ݻK.(,^wozc uSvǿ7d;^|ր4_-OZqᐗu49Oz 3ehu 'n)|4qE)k:8 `@Y2#GM{ D3hh߷*Uttj34/}_ΔJ_3МGOAAv):i0/gm_}v58 V0Y)a +i&%v>ݙtnnQ+Oܷ3L pl>]tfNCa5R8`fr5,#csd(^(bcɄKax)WNEzs`{ 3Z(ͭF,v$)'ZQcPkU&5dq(V`iiާVM }ScSkõ{zAY|A07[.o le%rZ)J+H̩ԕ7fQ@K8HCv%H-3s M0vg&psY\L{gKY4 }h1fQADwj.Ҏx5g!+U^j5Cf{ Hu PHpR<1_sC$M>X3I:U# ܪ2j줖nSSVo!&_Y&3alVnbY4>/ҰַZal!,AKjv,ٷǾ]5Z_}VZ+ 4C:d,00C,Yah + Q*/ن({Ý͉4 b)fNZ:ؕ6W/5EYgCw V]e:uT})Z]Q؋m3\Aa7:+"g.N>&೰{H pX?hhhx>/C-qt.NnߨiS׸@I' TqNV8 |]/U%u4;e  _.|p}u>C=ssn&q,Ą^e in vuҮdHڍc2RN!BM}L(O]1e4g"EXtݢepT$S]Zv 5hH ~.E:٥KQa_隤( vR,P0PvgnQ88ܡ9fMrenD#xNYI5); Eڹ9$v4 3hnPJ iw8\*F4]H`GqүoR]VX!6};3V }:ZҲoR@a%+:fEc'cuԈ<9q: gBb(@"N}6׾kmLzib\| bRfv)w (,A2~ʵpqߣC ,^3}ӑN8sbt&b[ժp{@;UzX۔,mTx AV++e qQW 9@"bx&C-i*""d3$q97"S`\νQGv9ߵ4͊aD 菉b`Bڝ+덒B9*tdҎ.!BM#0䧞y~; 1;SHIygƔo#4s&G 6ذ#~xCXn15w\y߾~,IVszm/0fݻ}ia>ğ| ąs$wX|_n#ѐw)~v&떨8 p(c ?5w>y]?Yrol<ӫ_~O&}xƿ_yÕیop E˄? "vj[&{[838o~$)7gqs[:*:1;۔9QE~QZ)7>Ɣz3d'?ϗBz@%r< vvLEkr4 +w_TB_x213^>,/55 ;Nd"BM$ȑh 0;ii!6iN0]3ӒCم7•I)m ew *MSCZ}?V&ˌ'[1n5V[Z ehR0954MjrUZRzT Kִ%aKZ!nF! 3վx(W- -4 DM[m ׮i nqP gBK&ϥRG8?$sXo.K-ݒ)%Qf,7$ 7-V<$rPmQK|1XK&{a]+5!5hYM7ZTy`|qx.!)7inQ)׮6F$bR7L𘔛QUŤGt8FL*bRwR__5Q\#S4ڠ|gz8_!/F>}v==2-!iR&HFQDgKWTU}[~8rs^k'0\kl,v@BTy )%$ε{􁔈J] g{= oGׯV%ؖd4T]GCkz))Cc͖& CT]3͑Dr:״sM7%ݮ_5UXa_kJB ~:ơsM4}L %[mPުkI\\ qM͢Ht香fD%$fZ%ոk\Si5"R뮙e,.6 KrM|(SkRSq皾<״Ju}6fc{=̶"yLF<rcط ^h0 w0ؼ_vlޯ9,NXsWϯ]# nz*]jDc?|kl$JhpJBvQ/2f p1|f&c§WGb0.ȑF}~-vǽ/f8pAҁ ZT"r%;*V~IMW69 dz0 p XEQrKQm N(Ŋ.b:WĴ)h뇈6:kbEĈsYO=ߧ/λf#ۨ+ .GkӛXozLL:rk׽w$kO3CKC>^C%-ȍo}~y)&+$8$q 1yma)ɤyRH ewqj~:ބ*=KvUڳzYX,kRP)C_9=x1JfKlEyQӔ[*(RA=.C0Z׾||3ZaJz`4}\VQꚵ/v: Հa6{5n,C.8B ̷\w~a9OvĊ-67ThJ%'>xoZX@iY^oy^}LB&l)?}>>{ A/zXk,5800 1z| lOVb}˰gn_Ò%."v>dIڵzaaqat? <,| [jqRችA `9#7M/ ~ k Z O$VbTJWŸ<2Bh#,,5X4G։`$4)pf&HXDŽ33A!&ET" C;*)Bϳ 012qӧ?>Mƙlj4\'L[ &F8E(9~Ap0 \l~`XAke T]^o3 )' u!Ӝo›!C(F*" (sؤ8%wBa',S{ۉbAdXU=An=4hsaA%|Fѽj59*~Fg'$f|5 yqUqb1wWW;JHz`:/9p I/\+^34ߖ{A׌09a{s"ߩV@*bz~#ĈrcY]fgKoE  p3)  tKm}H),ED2},W 1=f!4a:SI73FR,D%ђDJiY>:eL!,%)ILRL*JtH-}R"hfBeMw%usүpW mGj6(09-w Pᩲ*Ո\)LRkmX d6eLJjrWB撓gW _sbd1Y5PIְ gFnbПk,DRHJ>92\J%\Mg!R\ ' X0Bm,^n>p%O_n;(E ϷˇlCxnnܑI7:PŖCHM֋G~XyzhwX/`\ǎfZd1h-jB IgH|-J'poE t_{5sxYzD^&bTEa$#<&ZIFhsϺ+) !b;X>B1oȾG0 /V1:w|L<7|'y'-=!`&q>A5Py~I;۩yKFW'4-^4*!2X|Gqɐ:Ox/AjE״TdR'ip&#(B񴽇)!1a8N(m9F/wRwbm/kXE\v /3[g<zAB0`B{nLb|(=<}K[M {_C>ŤM3riLz i5Z= b}ZaplErɹ U㨖/`UrqcəxQ:B6ST*&U'-Ws &\Êhݞ^;ın v1ݺFxKb9ʢEݧ}W8ഖh*&唐Vɧ6Jh&),A;tsh]⥋ ͰsdC4Π.jMdSSR6̩nkd쓘zqӡΜ6>ju!򶆊okx¶SA  ^q4X\SD1-[![9WuN!.:'ѝKU9LT_J8N1x%C. P%{Jv.Z./,f:]]P@V5,`5ˠ,D .',LRn:kmDD\c7=1bKԟ,_Zx/D\t[ϘcL\hS6Bg̥cYVT[A'Uu/k?Q ~YRTYvc~*;c Quʲزqleg$1BXuƲ3ư*_Lxr,JrrtbədLiNU9Xv\ɘwA(d>0\DQI2pN⊤g8U2i48VΨc ʁ*Z fwʱ\Nh:R="Rs#YqwN"!)q&@ugltYʕϜ73Lk!%){syȮ7)w %LHOڣt4(%Qy `"A4`3˸{J'L z~r&ϸy0{? 1S>?%?Ոo)Ktǔ65PmQ̉&#"a+h0,|tifX5ub =9vS)##mL(1$dfVa@j2J"EklNn0O7?&BpͬJ>?-<3R+-M(ŴH)9PEXc VbaN ~4.* a\JB 'u-<5o ICr_tg1B\1AAKDZ]Ǩ>kLL8b30˽Vɂ͹9(c׹ϘOܗ`LIJQ3͞92Ն?')?iZUM~j1TX]nR,d'r+xPd*%s:U"yR0rQo.2-[ޭz}퐋5wYKq9ޛd7>=ygjf3Ws[| ԦkX[$k>j6EnH$:YQLib%Dč"Y]Cw)b$ҐwAh#Ta$dإd dMl򃥼^ QTO%MFrmwTo|fn[w+4@-?\ܥv 7tHJWl$N 9BGL;/b CH{GGb1Jtˡ9RJhl9*q^#[85rJpi`9jnKp-+/*%Qq:Y*1Z\u^,Eyt\pW\_odR *BzQ ѓ9]jLkf}B vޫGJ5b85bֈ DwK˭2IQNbƴe_\ z=n~:7߰&7~ | 7$$&o$Lq)2{-# Qø Jy {WvvB*ˠ7J ]tVUnC97 HdSZ^r>DXN(gC^hD@cX4g4ZX_e8xvx od=as6F$p yht 6D}ԂjWU)JHąLR$*NV5$($ 11_+ }$\QFYn٫t>r9;PHyDG”!%IϑTSAR38aMҠRH˸ st[Ac]7PIAwZ Qr__|RIm(m+ N%#!_tdJbNצdsH`,}SFЩq譪=ru BiT5;[l b#b3˸U;b۱Lܚ$-ԌfíyLм[U$9u۶5#!-'&v#A-yy _њq4m.q#UHY x U EQN+=mMkj}Nf|7-uufsz+[6蠧S.Y+- ,ȠdgzDFJl2h^yHh w^fa?y"DNL*]TRަkm 9Ǽ\9kt&XO̓d)ufܶIW&b{zo}*4j'˳7: 㤺~?_,yG7ҤK}OrQ*)7aUbO~^[16[LJb]-&8!w{Jԛw_"óI9t7Na#AB`J'n A^ԩO~D`1Hm&4*F*P!s@?ߓS?_zwwy4?Kbɴ߭B.D ľD/D86tcdm7.+0l;a[lӰO ˫pEƻo}Q,J05N٩AMD=E6"*m8s3U{.y4n=u~rB R z\oǃ2W)ң~|CjQnn'򨲨}zjqKXgf0jJp䮖,D;>2IWbgr@~#Pa`Xr;4/A0< U}(vuN=t %CQأlq+|oM[V _rZ]j,s9p8x]ySz`/Ph.5U 7lA+&EiMttsh c>VW0}z~E6$"}²7:? )) @ Kd1Uߚ&5-:B@1kujAP׈(YBz͇Y.%i s){O^(8J|:S_r 67g pY-a|;^4ʥ E=6˅O!i\ʐKF3i&e~")4IևN]|(ddrW;)co}'۟>8-۷1>bePҁ;!qn7?x&ØNC T?r?*r@!BReQm@h;hd 4 Z9>c@ 7;Be:7yJq.j.U^{)i9s#' {tBhZt;hgתz=ިhYEvՍsico(cB[ "1#9}ͮ3vdu' *{LM'*iYo'a0,cJV&B`˜=&j| 3c e$prYj/{WnAyAueǏ +2`umF g2Ie&?1U8xUQwzK*~yy˲Zq)maXd%Si|y9_LkcoZ :.Iy]9M~%L5@<$|fT?p V<ҍԥqOSa5 9! )l, ivܞ/=$6jh~lr4XK]#_J 0 =9HPrB~T] AIVKyLC[?䦕|VVnZ74/53|9Ow# F%*?,1"4xKFmܢgZv՝BOOʮuPû^=PejYӼg4b6J@QrGv6.)8a#:T:AnGcu4Ckc"Ă荐舣!A248nV_92N;{&ma;P9ͫ`=x!ay+QEɄDp̈́7L2&=l< ܨ{4vzk`qut4Ӓ{;w:@x~7%< -#֗ǎw/!Y=!~]}̃\Fd',W$ؼ 2*Ĩ9SyV^WΉ;r(lbY:'j \8SgR[%n YnhKKr! Q$?8-м/媙?;˔ѵ9;PN!|xؾ](E#!eNQq,'p9F*W]|bQՠQࡁ*,!G8tvYc%+J.ހ _}Fr9 .ݻը |:w33yVɚ9peAO_9jfZ^kD&A~YJXXP!:4MV*ys^y*Lmw'FNΑޞ~Y^3U!e;[['9IKR*`rlOcHÛ 2SFݠ/ 88tb x:iS\P<4ύ`Q+R,eL&TQa̜΍V)vSF(pS[ +Q{ŭպڹ ۧܬM>LϘ7>㉗vb26? N]9[~,+17?s%q{w l)վɼox/)0Ha~^.rd y/sQ9Ž $`S(Hf5`a59A#,Hc;nP`w\ HNM+yrbݰJ:3'C.p *c0H 56W՞H0P JK\aX.@{a|kr_䎟:̑8(ۮШA!ץ[Zɢ3 du\Guo>J*J/"b572w!yWmEĄ=kjȬU3U2{*ק`jz /Tjv\7oڷĸpK&XJ0%b6w3s6ܟX[?gOTIU|¦/nnԛ|AڦMy70@IUp{U*E-ZAԋ$9j.\WNQ$Ual٧UU_4hI uP42y"RӁ6wg!T$Tla)vfB!^;DP3c'd{ G.E)2!;ejk9]h{+;nbTkWҨ}r s6!P:}c /dٰ1e?f>9d| &pId.O2~k$Z.[s߀WSN\SXiκ9, 0eQ ~z6lӜ4U-CM^ 3M\]&DZ`Rжz)&lbT(bu RHVѺ5\M DaE'bpa\|-=R>;(֮ 6QX?_,J6[TG*/+% /4M:=m@rcD-g1^"BS*Ci2;JQ+j/Yɽ&zJ^;+Q(98%2j%Ծ$ʋjuJ5O [hY'eK5^!&z9@Pk=~O:q]GP7fn"rT4'RZ &kNJ:bn9yzg6,gʅ ZM !wAS@֯>\/l8ccii.2sQجfb} r3ӝUђ0SFQM3l8%^ޘo~vm +v0Jg!9űjM7/;t6=5 y] *d6T #NÈ`vKй0\yB8*, iu3nC|^"vqԜpv;]rɨI9i.~_̬2}GޫYD7~t$+$b;_Oo?Fَ17wTS_k`^XTx6ޟ5Zas彃fe9kq\O>w\>g,yF=rVC ˈ \iGwe?;ۧJWzJB([S_Y=;EjD0eQhZ#Yzc^L:=j3ApwvFD i21 HFH9ә/ݟǨE]E]Tvaa_(-u:'xg1VZu2>N(D8uxcG&&34B#]ѸTB[r`D\61=HQ̵le15;!,^lr0F;W*1+*wj4ɍ 到1 )Zx5!&OܑT?:x{7mPT,.%r3+Ӕ!aDJ$#{'9+Uxε$:)ܹ|wE=%CZt;}eHVm XS\dlbBR}gƕ{Mw(V)UpMIߒo??<0-_>a|ƤuBY5&b}PX:OqVS[εξߎW߹[0|I.=y)dn|W 6nygI1"CHnkYvKj+cV<-O[h/2R:|%^ U%U ֬r$ xh! '=*8֠<ENJq">$srjqāAЂDj,?Y!qhz 2*1\1rj(%E5ÕX]mx(L2RΝv@ L $՞֓bHm쉋[  Nq|S"D4cW*Œ\S 5FXd/T¬U4̳m%rB*/d%8jRD\-ԂspD ^3lkͷ_9HJhh+p8$؂ ZP qi¹c\; (/ R8Kb6j3~֘R.R8Fc ~fogY_].tHnh`݌EѨ( OA٫}>?1 W}l~(ӴSs2~>9_?כN _1y vKA褎QG Lע-]jdBc[hsCŅ5Wэ΋yt^Vko?܏&WQȿYA|urLήl1YDe -N"^ 2Q%> m/%%-YcސNVBURSRKWwd XݼAfCzJX׼ckQ0Ud5qֿ5Ͳc6x b̺lκggYT[+j˨}J'#*-k`FB&%0U'Vs<~`q%5 k1S Cz5"N/V :lQӝAg(oص_RvS)3e<9hPZ)djBZP)džf1C׬8Uy΀Sv4(#գ̄iM$pelD\wє)qhҌ0{q9Ҋq^ ى8o(MTbF ^,&3gJI,d&LuЧBT }ӄa*2 yfmXDSD ʔ&KqU( I>*VcK.e8#tVrZF oLx˜pRL`;pBe)@#3""XCŜa, b;hܝmq6i;mGr0^ S_b5X<539f蚼ȯC_nکPdIp\==gP°&R{-9/vzWM,D*._0HOy f=NHn3;7`O[ d@^#B[83)F2:*!gJvBMݛ/O:bb̎.nC K©-#~x]rW " Sxqfmfva/ @PW *ݔXb*w`8WWwy:3Q^fWqG8 lmtJ(Gb _@甞Si?a_J |n p|XXg@SpO,Du`Z[Z] z:I@kZG!E\Fe>Nl,w[8@}o:@u Eځ*6u=@;,, ʬ2Lrc};+NzM%:* $w ͦC@SN!㗜W_.%fm(#&%ol}wy $Z7u<+ ga#iRM`?ٹE~ >r.|TT ݪ7*-^X[-!!Da m']+7 P ڕtS]ObJ눓Y+܂] EI8,]Sk#t1TqL,nuP..-\.@G3D<D@ο1”\.bagmvGQn}iYp "mGҢX1*Zp;V]@%;HH1uwQefpghFnXgo|d|6{|mXNP@)x=.#)6oOvYƸ^j>FUlHC]V*+ 8z~>[5Cdch少Ʀ|HKh=I-.VaV7Ӫ6@s0\i[IFvx1 ;N^lD`3$ϲT @G"`Ĵ P觠prw)t  F@s0 V1Z2?zQϾOxy#EF=q rX.Ch)ب7,^5jHt0x(wbݕa]Dq.4zqw#P2ipṢ7U  tL]HVI=K5ܟ@aK5/Ef6TLeǀd,!#]p1}N":䖭jGZx)u#IuХ/vQ!7lv[RM/Mh*/6;oʓvQ#Y.;:%mT뻬h~luCU&~3#a暈IH gF)kHb0 7~!^w)^{)fbHc]:ˍhp`ra={ogo] `P6G /z$77ǶF "W1 q$6?%NP'kn`  E9W $oߍ^Q"-ŷl60 H`߸0 L0X ~vp"hD# #PKR 9NI0OX'<`N).aOm9UDxUgxvs!~>BlTo_ʸhqdYѶ 0mo|Tlc#=(=rDKţ/_f;J$pW_!cu@(lD@Mn3+@9ѧaaK>'\-Jk㡙ؽ"J(jAگ+p+ד(B#}Y¯}dY, <ÆeT@BUˉVocX|\}|z.~oΌ}c˱#$/R_#B1$8`) XȄ1.@ҩ $q:J mW碇O\=D *@a,H+IyN2#Q0@Xy?K)"|4rp@M*z " T+dr PrIXςJ 2!X;JY?!bZ%ێ1)nh1;R ``4Bm\Tf~hWoBUaܙ.F=ٟ~|s@\D #]p84Z#nN_/ k(]+B?fBn~.k~K'UF+.|VK}^r kIuyq<:LFǸ%..vqu1.Y@V&uM"q$RXqbOdwc5)BLD"EF ~qҼ!9s2gtnCW a1 Xg`t`֬˼<^i+[7+:ک*~ {҅n.}=10LΟ[?8p]J}'55'Mr&Eqyn{XCxdh$ACtHn{^te}{\ݲgo֪!aev!&mY B IRPjcL 0ӐmoZ)=U֜jն[eTl.cf88JX t O$H% bƖێҶ?c9 #†1jXZap,!0in=. XK71Uk2ITfv&kid!M2i tl}tg3@A${mϮz^^Εm9MܦZ?=HV-2߿y wBDWaEVc3-bb#G!?Zg-1a2~ G;LbxqZ?=]m ?!gDH!pޖxZJa6 Gl*\Pwmx 00i381 ;FhY{Vw2l10n3uN"OЙ"m]5_Z^?% |"ǹ l+3͌80aykݏ FUH3nwgBS@f]"bPJ5 oLo MuҨc?ճQ2W9IO6 3}.qd!5W7BrTsE?Na/TC$brXO/n0 m S6N?wj0RL0kN[JB =yÉҹ*NW#1ߩAhdU_ O]ڟKjM[ó+C֗ חz=|֗rHs\aݿ|bC:6t7tcέN_ԔB5nYAn٩GiXqd('Zā[v ,?_۫6_rKw9'62KVa+ taK Eu=}Xe/ua|M+&/lHWK3K᫨(D&GU%1,+xCvC^K̾ﵺK JaaWJ`irZPo}{<"A?8OZ (uua۔S|^gv;Ҏ#}S/{ּP;?srJ7/CKe {TH{K:]e'"̀^FFGyY } UFBNt}OГ9z|t<)ٓBy4zb דǎ ^O dΞy:*{]x$kO|\9կ3&8="j8W@"( D:}]6۹ Hj*ڢr8aJmU{*A e#"mTP- p T8uBzdAj5Yޛ~WhUt0M (0 KO ߱ ">d=1:7ch !%C B$HxlzNpԄ@SsIc3B%1P(/ѿU"`$F)C ʿO'Vm{6hXS57{Ӟ4 J4+\Պ%~3]~j8XL\ -y_zN<{NX07* *8-Ȅ Lǻ7Ws>0lū.p9<;יOjl˱+~5t·(Mvի$V%*#Y,5 `(dX0}! eR⹌w+ۇßypŐIJ Q,$C ɐ!qpiL gN&6163ﱁ+V(laF5 劣 C1$X|:f@hB).*U)ngP՘⨝ ə{ՏossjuNf`$ A* 5g֮4IR@ۙ/I&U BvBػ6ndWX|IeS#~QN)>n9N*/90͵D)$g+48΅c.ٲD_4R)36$k  a\l׹z8/Ǧ=%$XMSƨQ<>Q`⾌S>bQVJۏ+Y䏋~} }ϷO/&)<ax 06~ ""C4[Ob_x:/Ϗk6gW|*?skfƓC3Kq2q[c[bCd ӃT\H:'3wަsT ^H0)sYAr<c} TC3T mTR֭vk@ixa` 7nzx9cÿ/n ɿp$ Ƴ}>\ ̙ oP!Rn: +f&v}fcbaGim3naØL)ng: ԓg>j6dIyNLkSPC6\RzWfbNxHÔ " 17-]+MqjPiQ2}|O8BJ{.U:$ *'V.nd"(a<\zs њhZ$H4[tNZyut;),^Q{n b QDLl(t%Ls]g&zBiyaZ)I@@͹7g\\ /Y& @,ۋ/*'LJ'塲Y=ZL=8sd(O4\iRzoY͐̇S!+\yO=f9,|~w0=̿D_4o3;IloSi-SiJN_*?L հ\XUqZuZ7}Ch{(BAjb ދ֝C{/!̕%=XŵFK-n0LjЫ褆) qڲWO+ni"1#"σ5 S^;/a7$6ȸ9VհEG vd73uF-)ztfP?NC:>n:ψ <3F9H\rA Z 6Nࡁ4࠴A"5 |_W7}QF;_K7 { bp?N/zn^9N?vP3tbJ;%~ @ $YoIqE={TVeyN0mU:SB@Rx, B9fք/p#A'39jA0V <qME=K{-Z1,?߯15P3 \~H|ݾA%Ǩk{f\BݡqxPcD3ab!W 9aUV[ÜUjQ*  D@F ȃXb !yp&X+=U e=I({* U+`0b[3e4qFF{2Lkz?.8(OxeS?פߐ7?>}>,5G<  !j?_†6K-l)ΊLU~?'هgdw 3B(rpwy#+矷|bw0(? CJ)\!?~ ՞z kZZq"y8NIېFb6 IRVcVNj Y]J *.ǫKIћ+] X0\mu%6x`QA*εSR`d~ {O$k+B)nQ%ꈿ/~|:YYR upCר h7Ml>ra4g$Ze|6xy.KasYmnD"#Ņp*hrZ,NR;_).A@{$x1~<JNYԔsM҆22գZ>L6H%ߔ:$ԣt]r|P02lP0&kmV0O SMXTL'_ ;].q/q Of(zܚ[ t_]^Ҹȭ!{T>8 Ǚ՞[r9 0"T( z.ȽƼ%z-m&U"K^P+mGJSؒĤ}8m\IyfaX ݚ| Zgl4 nGwW"Sn-FY]C3FB s YJYl.=/(7~ڈz%ݹBk:$EHq93siCKm?&N8}:g.E't;w58,Ux@ι't#8 p{mF c1Ic㈌:GqHZB͆nhO5eOU[9-8~'I{淪/(2r r*wpra+ĩkQgҪB\X=rЙOakV?qOU?9#@YD}ߤ(J'ڷIw.>M~ ;(=M('{ȏc{ZZ3Ѝ? ڱ)7UaK/ԣ ;CҒPKaW5kAMթnQwlfu"DS. r: N46'+:7s5ful԰PH(eufFQ<\//2 _ ]6'~??+56$,H[MA_Bx on{-Ŀ}CvU<a_+ѫ&[T`sMU*l^ySe!ح/A~^u#DWߩcI>BZ(Mf/WkHm%_s$%TR3@:20t53 f8D orXVX ZJI54(epwtlcq#;A7'JaNswDAacRKf!Gazpk!FB6u0:ֶ/ڡ+z1H/ҩ_$O͆y/.BN S ?LI6N>B({aB:Y`7Akj-A%Oht ڭdz s DTT-Y`DJrW4QќhN &H3w_sEj\۹dۇY >\10'_3[~[!l]#STwL>z9u\B7?v叝RAuexFWƳesk拰2^3اGĕ@ZU<<*wS^z@]gLp# 4B0rIԕBIkJ UrS&T/s Mm䢵߻'*K/8LK}#KJ+bL c7<\0 yLx}O=섵#Z[2pGk5I;䊤} ^q_q6y_"/>T w/&ׇJCUN 'Xǚ!]^ZR\ₒ+ּ&Oơ /EdSټleS eSټle%2c3R Έx36ƭQ2VjXJhDnn_spf$|EQںw]{)#O**7EU5úD&Hܸ˺ZOFSe5AuNQ˼ƂM)L,UǠr0\2]S =WEbu`\,RVK5M3×4Yc+SW8͟vx^4҆^Gk  .4Hrq)K bgX=لey.o;#q[yGMu#5?rU+LǏZr*qv:'L06p99 rLw1RAf♏9~Ԋ՗^mZkY$:辕̆~XA+N&c~%ہxy4>Xێf%}Aī}z]zsz;Vio/<܍$?^MiPq >sp>g_´P6Gg|_ׯooeNG]%'m6M_`4s^ [P.u<]Vrk@ cҠU-tVQp Z¡M4!JymAd؎aA"aR1 F)#6KM뾟-]n49JxZ]0H%ouQQuQ=l[@w0wQ2pƣlN"4dvv-^ܧw{Zؿ7}'!cJno\"G \VKCIz!Rڠэ!b .;0##U7F!Wy?f}-_lxِ3HVTZo>X_xxj ?ΞohQ>֛~K9OMab}DbD7&[o/[6}ތ_ۯOeeO/ȵFox[zIlmx[ЛѰ Z\j&zU엍7~mo |n6of$U- ?f!tv ᳜UpAt''i!\HzOH0hDybBW7Ơ: Q8[4v Ya{c?Pۿ(R~rt??>E/`J&i܏{IDTzWp$ޥ(-&Ϸdba0*1 )b'k/כiӎF_5s'Q:JHYIE,"(Ԅs886/C 8-eJr"g'|S8rNu\;.#N BDTXCt]1zf&Pjmr*ކTǻo 5j.=U:yB?e ͊h0±Q눥8zA'}mőRH1ۜA8-Bmz)ozp}!"^x'RY[ dڭ\Tdp3z:o4l>0ق ǵR`-[h_ܓ_`|2~Dt1&&'֠9GSUCS*rM!)6ئ{UXZК.%jo]XA0YW/=w0+[`\֍i*zRW OP0,M .A=;E%!v I QZ:$D:cJvc#u$QwvП'85Qij(,x(T'] f+ 5=̥XkK5E3trטQ,pX!),U& Kd,3xZ)Gy5 % -p. UAYT]) TlsP6:1C)H%XT[iKFHHIlIfHYydr I*(Qů#t߲Cʑ[-|qH)jSV`M0X"*#嶅,U\љ˻=t"l!XL}D!)÷{hrC V?D†70CV#ḥ1ֈ#=cUJHj[CmQmJ(9p\|d|aK?!^*.]*~敇]aԊѺa%Ebtz;^2Gp&G(kl0tebf"Rd?ߴ_ `dr3ɋ7采ZbXwR4WB|i*YM0 fy!EFR^pCaHP; fY-_ ʓϰ8$l)qbl_ib"0$>ҚD*7#kd\xεČxSH x\kH !6Iu@' Ka<44|K#Gf;.U(E1AHX_Be¨upBUDC8E Oo Y QՔT g@b ZƱӱ`㘰( ~"K#iR>@%Uק?e(c(,jZb\'sy>7~2L6 9X ^ WhOP;ٺ|SiںO7&~Q [y X0< /G{%Vu5EWir +G|Vpv(a^bs)I !<3t{Wn]5G;{ 3Gw7eX λXSdcdƅDJ=WIٓl|`mTD-hr-й\V,57/yg$L_?4(RF4-ط7ad4vdu| eNE0-O RQP2dFR_"a}ŔLL*pɶDnqcI/9Vv/ij` I0HJ֘4e뷛,R"%n`IuuW׭bNJ=f`$D'hS~kr0.VyN,T,Z۬P@f]7<uSMM÷A -M%IötFnd9fqT,Tnet4a'F 1A&lugnv@$o~}-K@sWk4]̒b5W"fG^2Woee[ejd\hWb-K@ XG}MlSSY}o˨" n^_iI^l# )-_FؠYIVK54Ș"_﮼|x c}NqLhE(krхz6/_TM/צ@f,8r#O{fi:o\rq?a;].PFwyj]1H^W]Dt_iʛ誙E.滫)<ݝxwoŐ)3:QkkhY[%5)Fqrmuڇ|wr Ō 2k,WD'w3#!'1BYݔ.8e  yP ,H8ƵiyV^= WT­(]͂uX$D]yw;L0(.Kgn/ bʇagI6"W&~wvE!u:;yDr (wA$04KWr¥eT(,E秹 e!]ZgD\w30d VТG /y,rřuy."AVEQ]ֽP0lA둯,#XkFp(GR8yכr.I!d2OF=nV|r3\vq:t ČX1ˊt V/̆+4QhC@@|KNClH5X0Gתk۷-naapR\7や@] !9@5La#VlH8 $> Jl;KqB &KLH|Ȯ( vIBmW@`84@()P0!ϟ߯fOo]Ȧ cbVvaa8h_o%osL@`[~J&a<r3B_)DwQZ{Lgƌr(eLeeq-[Ů7YҰ[V#`NrV:+JG¶&!RARM8K]GжMg.[%A,BH!=ǭ<!P$n?刜>&x0G;fgWD{ N$O8Z:'_M@L"$#< U @kD$ GU/8m E5r3s',uLXeAl āt-% <:LMP2Hˍ;%|l b.Wө1d?}}нJou[[.)RD2y&{lLOA%.Γ[M\{v5]N+HiX$ *xf n_3I|Id5V!@i?_[:/.NTbcv!$|!zXL8vC ]TTģ=^uXxiö.sgS} vQ.tPkceŐ&O-Qu@ǶWw"/.4 rf ͍Pa wB4]_mAQ2Y]FXԷ e( 8!f9rdkl'pΔlic,(!oi6T`ILZ#W.KQLlF5B<`k U1R22pE7ŽiYjSw*gف~&1Êm?xԧm*;11"<;TA.0st.mCP\)}Ft'.[_(v Sj;_)Tpڏ|6}0$k3W QJ=q̴SG| |eC.Т9قRY/u@.WadF])[9O-[z !e7h/U;@ݦݺ4N2d$8D2@9m1rqFCRo^PN5gRd'Ө8٨fU+a5Zm Nʧ~KGdABV/ˋ"Z=&*1s^BZ9w! ADvDa)! @[9t%f0qp W2mw:pߞvFtYNs2}r-lavњw%݄=N8"Bղ(׍su&t$f7&'F:rŴcK DmGZڹc5@4`%r&4*pߴu8  $t_C:Qp>zY$; 2zZ@w\B>e:/V}bO&0.ϛY^WߣD$G-Bimg7K1^fXm93`G2S Da^yY8zq\lAF)%P4i% An;.&fg2D tAr ^ BP2+}%v͖h`I>H'S}):a<_Г9UyR0mel`ҍXSr\_t cnͤ/ɥL4.؀}pN-k%dCT?/ڃDhMz֨ EA..&%=($8epjEr8"q!y-l0'qq򅲯i<^'&`Kv,hPK]r 4JUp9'JԞW΍@J,󮀣x66 \RVjnקA{9&z %=(zOAHIO(6|shH\IL,Pɜ~]sbDh$ozwUNfqtp7o<_He޴Up ֍,Ij2N *ű-o;K6%sP30;e:{=[g=vPɟ54"s@?{S{8W_ ot@%rdb6F0BC! mNF"|5~> BL]D@DRmQv9:-b>oJ맢y{p sd=KF|3%̻V(ѡu}2uij}AX4 #D{nofu@t0M jX(Wղ #Om0㑃.x(#v(D e N?v_) ;*;hHAC%\{JRnO:aeȓJkcnqa}%8LCbGFaS rx85ۯFY(# mtkwgKpRZvHfEj% ?N:Ѝf]BC]rz3iXjSH{ ꖘC0 $.*ը:m6f19/CPMo^ {hDp1c"dL  &Ω(XA8 )! Lmp i{/7,ΤϿ/o'x񬘸?QkHL2.9JzxG%jʛega%ge~V{RntNUw$O{bTW+C%m%&Az9XebJ ƛ,ɯ7///zҌC*F&UrƧWa+gU5riwv7xzЙ>{эoo/f1W_8"7K$RIm7r>Pys66z:{RKi2z ğśU-6r]j ty{s?x+7ډKW}" %QVn:ܣd4^/w&\g^t coɯfqkv |Ёl|S|6 ߬fU&~**Y+q~Q3K&+ߤQ^s8l$r=Dl妼P J\>ER?KRoV*uR#g-αOFBpD&0I03뢙z. TdJeͳI45[DOjk]-m%e/,2ŸĐRuշeH1(K7왑$ T ¥vWJd,m?w \G*\T~( @pwTÌe^!Y]H˙ӎKB]ScMGCۭ8*ER8'],mNh OEH#VHDͳk#5뻏uqXQ*$X&X31L{g`n9\ƈ`8uNX3jY!e4>ca@)"U8WLz~6_?>Z/Y8\/^_ɿs@xea9x=,V vgM2Q& '+!(5qNZj'90(cJh  /D<V*-ЋJl!RN=rԴN&uZm } =4٪%Uw@Ugm ՛j; BMj;O18' a O+ Qiaa1>V~D%Diwg#z{'MY~QZw_']KpŖaochMҊcIx~ưԒvD-=RC 4=Bvl,$y@ Q, @7C)p^i'K!&X(Ligcn0Pe&LR%~EQͩ54/Z}P4ᇦ*nX@wt)B qK8$m+'*kƤU7$W+ԵgusM'7oҝo(1SрE X1;MoNڨRDֶIlZ}h9_LTVi ;uѳȹ1#L{EgÖeE֘'vC?u 4ZAs)aT[8 ^o@*4RQr[S>ZЗ\]݃BZx_uɀbHFL F"gHY4caϔ11XH1!E F6pI`!PTa@B`%F[l$1BbmJ+0?K.Nd2iwn>F G;!An2t'W3{-ER# IAk)" xQ< @>Rǃ?RHA!<`H1pWF`ǵahA̪4%eBHSfGBU|V"QE15?&2P%p+b,N%" 2l W(˝}DljB{hsBK,Sf TjxG ᥐS4>v%r8Ɏݕϗ, 1z !_EvR , Q¬ g{tFO-uXlG;WR;Wd n\s*6Vђ[&pjgs+dzͭDǭv[l>-FK#gЃ)@,\A@%-5hxYG4@)W`Gl+qD @Zquܬ=k tG|]7ߞ6`l&̵2^a < fߙśmyYCeU2|IrwXD)>\ax.?SXN*+s4$A-ѭ /_g|[F]͗:A'muK: ꔄ~ jzW?l H!@D-9&u*X) Bltpꔄ=.K# Qk|["$B}V OHԅDEƶi%;>V'5!Gw8'EDYZ }uEVkGm]7"/}=)$_zҵ,Lk[=ڲ|DG]%PLHD:d& :A' vsPM٥W8ZC3Ea(IRK˰LkqRA?g?mTPc~~2Gnq@>IiUxo%5o u#D<9c`4 D=ӃQj5sPXE`YtD%,itGg ñq9bKd/]#ՠ%l/뭳V`3?^UX.0G=U_F vSX˟ꩣ@h-><쩖5𪗊xxq=նUN"8I%h[fbzܞ/ o,GQL\8EÕC]Qc;T;K0X!1|RGcRb1J,el hdؚbGF`$(ul[ޞ"˭~RB\+mE`0#+UŠ5xBDSV{F U^*fOu7KՉ3vjk _z]츢f:pDB4S *pwe[f%{>n|o pD&|ZխmO39z=$-w[s1AA[SOoX(wlXŦRjt5)O[vRLI1i'ŤMa:Z(JޞWiרNNu6hc"ȾR$F2kjɁr++GE~r9zЖKEyBWKW8ih%LCt9urEL[X2b}+t71Y/oBmGjlËtl|H^rm29WNkd[g6*`Y2-ICw{\e Ge/dH|_H"L:G*'4CF ml[:wzj;zX| a$-{V\}EyOV1Ѿ'eуrG;waJܵtWu߁$Z|zO6 aO&gՉ$ ֌; b+;rYWh;^0ϠZŨ#Bh2Q 9GlT6dr9jJ%yH29 ͒ȗYuKח'%q~=)fg+p`~Ӄx=Mo5C# X+>?~~$σij>i9>_?rW ?}s4#'#s R%&Zp4nG6Rr7x8\;Y+iBժ VeR0Eko|_? +ƴ1eJÑ(8t+w˔+zrRV夽wH&mB!G`ҡp9fofWęcNpw}Q2}_4[m)pϭ*sL*xZ2g<[0lNBr:eK}S:BoE = Dۺ"rϭ5!$PLdAHK\yF+V ̆Ҳ!<["{|#CH#C5˒ ٔf!dd~ 1PdPl9hNig&c ^:Y:A;6ֵX*h#Yb|Ȟ^E%%:̋z K%2@,Y) VԀLQj1)uP4^B{TE $Vz)P9)HkML1IC(vC9b_Bg"2y1"zb'В i|rK1Y/sERyr؜#qtz_-\@ged>}i5l ْ3,꿜=q$RVK`dY %hi9%c12{)س=0k"D:ɣQw/ux'*+ͿZCC8U2W>@Ȝ0ABygEvŘE, :5bƗ%z[D'R/ +12 ^<.̰J2fGɍprBɒkȸӖ}``qXM7W?iV`(fXŖ!LRˀi4du+l, (Ga#@J<*JJ`nˀv՘qy )–n: nTH=]7zsOYri0 7-^ۡr -qF=QHkn-ٕB|MBcPK@n/,ŋCABGRʽ 0\P aKM0STu={w;0DϨʊ .xlZsr[z;>Swgp5{_t, ϋ ]f ^>&ywӴ[IE",7o.;w܆fm= limɷkh;:|?\ﳉEpn UK}~7 ~C$ =x54d"p;n<5[`}Bo<ڌGhgfo6⍔FS=O7@C:~pWN`U9"VLVŝQ.>A > 9"h:S DӃ],b>P#ywAqm81-A|H%W]9Bȥ| +փN 4':ɀ޹-,8yҟ]sdR׋8%r*+-$KMM=i] {HEHl>&K@<(G0Dc~{~jpN&ڥ'okdߎI#f>Ɠi\\.gŸfyУ;CGq&Y,>ѸdBZI__߹"E}KJ;x9ԐS n*?5[VTWBGi ^pH1E9t0(o3sBbf? q3G`MrmZLt؛W |Gx] ,@R)}?Y_JVIgu:*0g\Ph :`/ZVJ|]Vj rɛLMBȍz&wܨipΠI*;6+Os:Y8ӉCQ@jx"C:V mEFJd;SPnd[UPܗ[.I&l< ͹,1Ktoh V.& 诃ɤ&}x\oyv7&nkdM~x(78-k+7=+5Bp̉^1 Yɋ'E}qX 걁-$f|Т z><r4l4J3%&j2נ-C4F!(|9W>Dj?Pj{d2Rd vȵΨ/Br`ZZHc iq ɽ4 6Z,tPS,uۋ &ɘԲ߹w@ ұ@:_X.Q/}S7ryAC|GrY3agmZ|/9R= $H!J*څvGE"ʹiO+!3d;O^{*݈('\7ů_UNl,ɩ^qhz,Az5eu4lxȻ@7F+swFۃ;_r|OW3݄#Y.rge(]Ny9Q"ŤփPg-Ndx:k+qXen%UbVj)@{bd!  8p^ T(\ Avg%Jp.ˑQCahYg Iќa+N,ukiwsȁ/^' VwΟd37UvΧNn5[nɺWeZr nxn "gV]4ncT4lF!m ۏᗳߤ}Zt0"/?|jºrz$̍])@3I$u}9^5/A. ʄCx: (h H`/nz<W kF8yxA%GlXXBλ+ DrR:;gMF1)d>.p Hk~oon>ɒ Au<R镥3Im4pi<@k‘XG>ǔz`Pki0C0s;W[vL |x-gbʳC3="?#)v(#L3uaF~9-#J` [e1"=[n )vu5)Ȭ2ҭD\avk1}u Vx[u‘ICRS4Na`͔"ǒl3\f n<9U 90eUOj\. Yv0 |NzO0R(@a.ki8DWDn0vSM)-Ƴ4;t8-\I߲2GIi`tƆq턡9)ao)L(J)^JWTʢ^JW*a P^< pVT13#QؤȰlçc1RhJdkZELb }4Vwk|֤WZ&cO?MmdzL`Ph,7[+E`Zڝi<µ50v RHP 'ޠLs% 7aqt<(D5LDQ oFngtM Ԇ%#n=@M5T Gkb P%=L@ёguˇ|J_} $N" :fPj0ј;1*TÄό@cL1c81{:YoB7) ezìQU!f4B(/ -K  8l@ddm*m띻 B;x.JHc{S 1XSwV#pBe7ηߋo,,KntcF)8;=ŭ!XPbO43 /FR!w6XaE0J{_ L{"gE 0]1 L+gK,سnJܻuri}F1mvDƪ-NP;|w @Yugp0$&{Y [ (qԸn+ۋO"mRюh{qZIOqb.ALwNL-=ׂ/Uun2l.Pك){K|ud{J3{l>^rT2!Ԓ9[ 4,=M'{ZG *ٍ}B3*3vnuyzFxQܡjxXmQV[?iڨY_̶1Iw# NwJfo ultNYv#ʾԅJyzcpG[ݫ}vF#X1,Ql޷\}uò,!eQC7, ?uAVTO[g{ʊV0)62¶Ù0NvOQN[fҭ[6H[4% CwIφ My(l0\ 4F3rØCjiջ.4z]> m ݝS~2td^جǀTfEzg/k{cpoYlil(%VpaN2R*Q$+:uaYmCL[z.tdȜ0&yɐϿJW yYw:wdȜG'Fsܵ^dzK#n\\L\;K:|஽bDѴȡȨZ^#uzX=V =Ӷ\oWIt\kdɍDM}Pr0!Ȩ7#Ti.~hfa~N6JRb- ̂Sa߯IК,^tz(~AicC--DSFgQ#ӕ4UFZ[_354mz-P>chN o7.DGJ"n9zDP-(2&X紶ak"p `aʈm 6DPL3\0L+ܛ u4:sN)3%(P. Z ss@Z%U6R X;QD]dIP XpJs@ ȎzqڤK唉V 5DX!$K Xe`F@ {S 84o0T袲mPdlC]mT76 ?[+|mCMm 37HNjf&}]WLl_'՟+fWj\MI2P ?SOaFփKF A'ADvZCp 1 Z עkg@a.BM= ,ԊsPn{ƒOyH-T +rutc&JV/nyEX8XV0\]aOogUj̥a1Œ<۹KZop.YTȱԂk$ JolDyM࿆Xv>׈5>0mvc4@;ۻrw~ >-!WLt3.~NޥkXY|Q0$6I88Q\.>92W[i߽*i Uh(霣Ƿ0@1yŗlAxM%qgO,<6jQC?>ރד ٬RuȞ[&fo AIU?3fpt1cdF|'oq]y;Ŕ%3gs6l/)[wjd^D /w_f=(f/I x{eaKDIe{Tc<@z#Me 둓0wZ;_4/Z(ΜU C}Ѯ@%I;h3|YU|l\}1Gʾ. Bt-|sX%31_p>,ncGU<9M ] ꃏ޽B`5ߴq=k(eS8",;"BP;.α8Og(ٞ>9=:M5)&BP{*p4*ܷS((mVc9a*Bƫ! U.B@r`UՌxK囷|fwS~x.b/xU_g ?_YʓnU<*su>j\&<YVZL:!j[31*BO.Y]}1VWIGW K$x=-kc麟7CT轋g@{{tYKa *pF0HE PrJs1Z"jX[wmmMm *=d:}8N*q"?= -BdH$.[@ח鞯gp>N޸|;kC'TY0}֜UmU)΍o曜6·nk,ؓ ~=Hm2Wy3t\0?YIJ@秠WxG O/خQ~w 0$Ỵ 宦ɷԓe, F(- ԼW$B;f[\NS5b:z_oܹ|Χ7j qt?:2sI:o-]?3V]/!jzWK]Lٷ? t>tl6|׫'D38()G  AJcgdY>3sw ͗seҫsz\gdnĥyʕ+W]XJXJ~#\xwey≕:3J s1yLtt{݅EG!y zseU rJZ%Ejrklj "Kot?-Npn$zM1˵IARV-1j`<#N9+/S&Z)R5DvhÝ!${kEeSƌFK{ &Q)qtQYfr`"H8l$FxFBQq`R-I1wuxc1}\+kiBtzq^S ޮ1Z_'PLB'"jVcτH6yZYCG-| q:. 뭆 tNs,ą-^8P?N+ |vXՑ :cnpr7g *9Q_Ƣi@'~fpV=L59)$/n{vSbڨO9KWy c I28@<\N ޜs=HP5D9[_?UNq,^׼#ѻuŠtmcGYiݵw<жޭ y*I=m'έy7`6ӵM"2%3v:wCC^v)^i ߗݺbFuw{4䕫h҈j޵^K$E̕}ZĈϵ^+ٜs`8*g犲\Qփ%hء?z'WQ}$&_\*U@:c=~5T5Nm{320 ̀i ;qyfK3[ګ%JLƵYZ >̬fB1k4wgf-3k}X`5N/Zfz|0kR̬k ȃc8c,3kYD2gf-3kZ޶95XWYR̬efWKPX&y^f^&>@fM02^-c$,3kY$w$=kYY ̬efWKPT9̬i 3rp̚T̬efGKЂ6y5կBB?N{;&}+6,qTrGZb>q[O`S0VL>OUmƂ7zm2L,oww30L宊>韤'\(tv4 ]XiG_<[¨ 9`D $p^qo|ڤB$Pa!ϘB -1[ Z8D G6pśi4o'!yR'5*-.TZh 1a) RK̔JCޮ&=u7ʶW`bH d0܂0oaˋR"pt;_ og2p3vO4p%7!j o흝x-ͬ1YXKQCLuu:8~op8x?$W]8 TT6R;GJ-#\,8"~{ctOtnj,{}eh167C8-TdZSaU *LaD#&VkY A $)9dTތ8>N[*9CCTtQA^\tYK4mPUQDB fIԐTNDDHy1,tCUQei#X*%S"& QZ$1಴bJ/5mZ:ʿLT*3 P;L$bQ/\ e<]ڬTa>^TS i$dywSS 'p0OO>{N(٧fW AD+,P_yRN?G# O[\w?̬4d 1ܟ:P҈(g vq}W!fBQ `NKx&GL , 2c(K{x?A 66(M7WI7~8k]+dP%J.2`E hehM/6O8^e(CyJ쒻kXBIw@@<JGil0ǔHiME>X! xTpHv_6#$:Oe2!¢ƃp9lN 2dFMGB;AgF]Xؔ;R.R hRJ4#\* AE*C"4c!grG13$X $ ,vPÇP a7D#+ <2!(vK'D! Fĩt>RKA$1Ɣ{/(15!B2# NJ͂B -!fyBhBS6(R%s l6կR&)z4X} Lrj :Ґ/`F @ 2E= "T=6;AI#eTcce'fBQ3+oTŻUUݜc"O*QuOЌw9qp`A|1XډxdVB s=ER9Rp::0dj r{Ƃ,^*dK0 >}|ZT2dMS6HI^'=Yh\75 AֲXds\7Eh ut+O:/Fm?$$O ;y( FuU: ]# JWh 7Ӡ)B$R#h ٴz \PiCOWsZЎ8o(c&oȹLanja/3%Afշ/qJc1뱟 ɂ?B q{ZzGxgZ!{ZJ,B:- P!0Xv;$f:-iڀ 1C[q&a}: |v/s3wvwxP݌R2 ܹT?z2[c)b6jh$\X 6;F $8h'bOvK&Y!˵A[kS[g@P3;kc1FpEkfN3Sю$. KWF9BQ| AL)bLAcq'-1~8=nӿq~v꾼9;i}E' ;^2 4.,+/)oj#m`5jd}CL~X ìgnɣ{Kn|N+Qm^ݛ.ؿE24Nir2كDA驃!즺K:Hg@5 !8mK3$pSL쩅ᵹ 姅K!a6r[/%Mz-YJf]o#rW}yCZ#7M>wk]O5cCIHNwկ-m}-ZĚ?? vkEuγ??7Cpݏ?xȝ?^~^jѐ 8@9vQh[Wҽъ:sT AGWN:F\-=$z42Mheh%w\?ߙ :РȂnu߭JWXDM\hsd$oָn%&+YR6U;ƸafS2se}.[\xh**RY'#`+N?s~c4*GQKGBHl/de}/r | ʮjqkXJ?U+;^$x8Oͭyǒr=w~X>rsq&>V3R%9\\;OJRf8]Y}8ribxi. {K,i]Z%g/o&r@Q`z:L펡Ҕt>@`Q'pVP\TѪ t:0ӡ#$'ush  kp&P0X| -DOudxCd5DV+ ϭSе-0u AQ3Bjԡ{P-!1.'$Wgu;L!2$_=]NT%f&/x ,βp;.f)Hf㖰`v `B=WR%P -$lہZm;x~mu0ģ m=_]b6VXv17'%D%8{(1vjNi˕t~rY_?Kd=. ~s`6XE#-St޶Aq6dh]eY *y]x܌܍yk^%%(0hs יM+HmLtQGn)DPlb6 4㧁W϶r?9hg8Т|3ϓj-ͮ(n`8bCzJedJ1S%_?;IAT{ $C 2#̹:d"(DaW =PLh*AMKAaL 3ƥvl$TMpY71 Z˃c"!%>[⃧Lߐ$Ҳ{ %IwwhN{ص&%D5li+d(O1Qd `| L}~B=jtR^p m02R];twru!m^i `3}Lώ*LY]J\PZ)OTD]lm4RvVlﵵz@6ťcs8;x@ J,s`E"vz(I0=e'<$Q#?IHP|\S2|M@͘1%6Ě ֫D< ,eq.4h/ӎ3*M`Y$g#ȀrpBH_m}*,8OX8Q̫FN6l%C贔pyDY%uBV!:<$PXKh 8 J3`a`h0Δcn[\ĮBk6Y%\ f㱖k ck;9 aךKBxbGϬm3F<́XFeEATi;|q@fEFrv)fFlg<?R(-3h$ \ (9 k Gg3[iBnĂzc׮[|]u;ӄk%si1YPoO8UQ %8b93IjeYlMX|;_k-6QXY[T(e>^/YS_d%=zfC֫M4Hw.vQhBǒxR٥A; Ѥ5t:J,hcdy~=EB%4$k*mp"NQOs.x>¢_Ba"pv|ㅤKhXԎ1^\ 9\s\Xn{\a 4iYFRqrpu\qU\ b>A4AzU~'>oh:~oF< ]Y K6xe^WyeA}@}2x*0=F,oW5h.3/&&i]tƅ8FBWڣliaH;8>vTCa'#f\zef x=.@"^pQDX J:$)GN0(Z&-w'5b#ݐeno[[nن‹fq L$cr"FX.szkSbtr Cɋ"As .Ä 4Bu!E4ir))zɡhdA 0E $cèȝT+V{XM#[ˑAbbC ;m RU= ]Li]XT:h_ʶzpJ ",.A8y6+joOzKg켨>l'7NpsUGcכ&dhW|A x1B_o/ űGOPX噑h<FU]Ւ5Vl:/zd ~=2bH]]+咍޴E۠(IQi?CRҡN1q6\$o[HGqNr =OLhjof8Uf{o}å?!oO .=7oOo.?E{!t>r~+P^HQQ;ʩԝy/jGsK]JAAuR6#Q-!7Mبꫂn%"@%ɸN2!'EUB.U Y%vQ1D+ю٪VL 0G ] xCFQ(S IHԃ-mw& uo.}| 43e;8!rN1ϡp0#Tqj{UΏe_${~ &˯kyS29M^^ǏղZݤ~ŷy7׈)gso:)%N>zz 7_ ٛ9 ,wfcɷ|89~BQD#LP+06"J9 hJC^iXFtnXĊ醭 ߝn؀Ljcq# rJs$#v}}1eP1O(֬$WCwdɃf*fs\:qqWgYeuzY]=__UV>(QRbi[ 1ES.: ȃΫ@;=P+:{ww'___-txuvw''SԶA%h|Xykfyh=YPK!_%3$pͅǓ1䦙^, )P2!IS#RPhF#D7J5IC@T ;(YI)^ghi&5(!JϴiWrJ*>ҚT/Q-EZ6* 1f#wu9d6&HhE} >g9"289cH9#9p.lg ''Q3Qg66j"?XYj#vV3e&F`$G4*I '$C (;m5b1A+'L!woU父J\[_l:gy:4ymxAgZ_73DC͞?`4ٻ6dW,p{W 8gd@HJt3IIR҈"̐#M 뮮=| 鼐%4F2 J3!T OhT68bhnRQvHr{Amr]G(PR*+0[ڎb`b%{JYpsymQ/#jO#]-]>ç"R"-j*yw~2 A /7 6\_f(d]ZIPJi>M)ar}~? 5|;$Rp$.`, Q``D=Ѭ~G2/l6W,3X9İR^_-.Z}M95S+\#Bu*UOg$2n,VX#-B,Eȟ!֎oֆ"e͵h &Z$%$1haZcv ҁJp67hQѡG=N)a(BMӥ[m`-s&jOcźAjvP>v!dw:(&Ky(8!(ڠC<[uγWf I$VCס 0P]WDarT2Wo ə~znBps0;_|M(,m% qu*~cp1Ec%sk R5#tz9Ym&9,4{=TJ}K`SwHo 1Q'&p؝˛zjO-XjwOqZ1#Fu^O٧* "P(yHM]٦BXof ]T'"@с'$< 3W0+AB`"e:i 'oϼD2=QU 1iJ|lc墥Nx. r{r9í -H!DNJ'g Pnmߗx:aZM9 h!zObK9orK) CNY~n 3dRb'%?n.QxKDJ]ES Y҂)7vˋm1Ɩ6" 2faWN1=ǯ'G,]fXwyx5&&}6TŚ~Zo7~CN])Z@LQsQQc^F/=7D+:DfbW$pTt(󂖜h߱eA83,gp!y9pSѣ'4\|ǜ8>R球& ,y! V:B-ȐGF6K(x fF:ITm2HpyJF.Ro#\ J$-e.vSV 尲T/aZ#|0T?¶D~R`! O0fC{ۭUM%>8JM( | O^[#iS?x6%Dd# (t&Kr򦄩TYx!3uՇ,oY.ZMOpxhn?߬c#q~R,yy߽h|{s:ø>_,>@>Y|}/{,)9 $K6BrtQpӇ ڈGp4^B+}#*C-ݒcѲ<1C\PUAk?.]L,?M[m'@lP/&}RGL >IǺvS*cRWƸוsҕ+wc9 roJ f/wf5Y}9zV*`;.+K`ىFolaO,m͘+;#xڐsd #,7 + B@ŴЦ^Ykke0ш^R.$S1U ը0AY&f|RNqڠ"c!0Aݐl31%ZVxO}wK.(~ԳӉhY Aanҏyl οۻ1?yȞES`F<׋4ϻx"*ֳusys~`P}*$Eo9/c(vJ XR(Xz7`OB+0ZU+tJ8w/O}~,8mTkpz!m-fOGASc#=[$Jɮ6TxGZrvz{ǏuB;uIQ!:n#{=6E؅(3EW"cIkC^.;ޡ4Yw#/;MNOn汘@ZdίMj@jQLZUȉȐvN7G [Q9 ʉ@jN>\_WeAa#F6EtܢáFCs7> -!9bDM&fb˻14(V};)t,K{s `UƸ߉9hǟ%ua9bx^31;k2I(2I(T4zJ("=Kf-)sv/YX'zC/(`9:2QA+SB$C1L¡b\[w0w ߷'vy]& %)&i'%d ytϜM77ow~[)ƣo,gxI9kOUQ~.dw/5+/}wr[!_N0#zw sd٢YSGeC~V"i8X/Dܼ66=qjzFߛ|} 1=~Jrf0N)ʔ(Sη[y?fI!ebLZPZW2%-x(SRZ%|Cn>p=hQQbi&yHDщivXZ{{:)5曒 =:FTQ5yK& DMf!&sӺTG,w1j- i3nxZ֑J"r6*%D>kIfYiZ ^\E 5DjFN2`bmk(5PѪA?Xg݁ Ehe*ﳚK~2kJk]-󏳞-?}B?yZ~wnJx^n?][6+_v~'A3L}lng|[[`)Jȶ<,v:N_dU_.d>]LySS+^^gCL]2zIKo ^@HԯIO&J~?$evNC'Z2F$v;2MNjZV5!!r#SVJu/H5M[֏MCLܸ%}V2"<=}w(,M Th4EGb8u+- q#B96":)QB"Й/|`pGwrʳR#UDWI#7p>lS  ELç']|XPdm#heP-zH!-pŌh_ww&*u)vQzs>Rst9{0'w;̎b{>p*1[I#CN)f mf-lSI[`)p& ^\4\}~{  @Y@+!`ʘ!B"IeE8d j^x b*14D:Z`i5'YW'˻~K5 >ۄx7_X猃"cm2%§WZ4 ÀQ+u 1'"x)EkaDZ1I#T.W1 ARB8YScI@vPUlPhG6rxa91*q]lQ}U* ěbo;Rזdus FI2>r!0 <ț{7Nb'ҽX97|3ZRS G_]T„U eMGLщRoȲݭ @'CIռWt,WឤBj2w؀>8%;la <M FTUp8юkE1y2Ua[MAp9Օh_\1! M(42J1R$8 8I鑭3jVO&\ja +9Y",f( =q*lh0q16v:ِP(,E4s|,ZTI*<ϿR,\.S p p <_!$<85HJ'`V9p6Y,rP % e 1=~2/~M^/8IS)̈́h|!F ]EPU̗`8 lU̗BĂI)+7kwc%uүF.-@x*^)+^V !|G9UC-|'@Rae5IIZ;I^~kF%3Z ) RTo 2֐>ԧҊ5W0 X%@ {zZׁ.5ebG+p++@F%D(X!HZUChH _J#pڂ6%R@+v*Hٝ#t;#,pɽݸx3 .ܷSZ(UTTUX<͔z Җ LsIٍ"CX!Jyegڢ6?9PLe|I0)O<81_L"Np)H0(AHRy%;d Hihm!t Y4F`Tu`FB*ՉI!% m'@J@ HRXǐ  M((昀wS`2&Nck m.T2S4ߋq +,)tLUw}X &45V(D݈nл1/g0S!Bn(WsM \ͫO7_ Y 8[(is M^^)AvN'ggJYJƽ?81M~'v4/٩Hzv{CC;2`?+@ CJ|FRpK0.d<I=R9?d4O` \iyb \0uf0I ߨ21KɊ~YOh6}I uKCh+4?xz,,wN"J7!QGxVWČ3&g (:qaY_ܴ_m|dVmyse7F0[mV.F:kNa- (1pl4ޕz^w?sUtVu; J8b⪛=\\<dr,,G"WPѧ ugvwiWnu*%Iix~SxjUW@A2arݾH5:#uyp; .5ikJ}j}V]7xLSŚY׬jp)Zw~K5h$&?^Q-0C:G}>VreeSy-P,s-PD&!qL46 S"Eq(eL*f f8c&;!HMNAeqUsTJ^dOAVƺt/ī?4O޽̊l JFY< `d QFt4Oerw)ZJRmRl%\s>˓AN뗺"Ӧ/`?, BʤaNbYB=' >L7; _2s GL?QÂE3YXZEJ1Kf6U}8zPi/i{>߇W^`7!1O&*bc!|$rDQL.X?R_;\Өx vir%Kh=}uޥb3{Nw?bsx CtmT˜{>` D os/vOa| WFS. Mk/[í8qubaaC:66{M\84 _5~$y3vւGOl kF'tPmL \yI\y!V-~+uG.+yi;3Qg4E^}>PCTYPZekw*0ݎVY2GןD'Ah gÉ[eFqUB9m]*iGٳpHIdjǸ8s\nnL%usՎϩ[+%yk|ZmIof+`fCBA88Ҙő2#$FDa$A㜉&3:mf/<uQ Q5`ޠR7I١|։agYvbMw㩷`:7N,{;"0S7 QR|̓,2mDtn~f\$JT:s1)[P~ׂ ,ce@nRg+j7L|rIY&~Hȟ\D)9&.{x76p$9y~@r9Jce̲\3~'t_y#7/fz37o0X]/w, Rב||ȩY*kzvYZ%V)HDqwE'Qcvg<](E~tS?MꑢuO6`> g.r Rx@qJPX{K11ZY^'~{$ öWZʶ*_VmpnRlm ";pr^g`.ZJdI);N!% oҌh!䰻uꪃmTnA~EMW;&I$[4 2vd]En͖BF|_2V)K>z+!y[V}0 DJ%v`wǿWˎqwhu`oL.*ڤ*"m&d޺n|=O4E:imp ~>OˆIfJA$cٞykHqrif;>wZexvA{4Kq)at*W8ި#1,~:mC[5M2eLxdEw8D2Y(,M^JmUZ)KV R`bFdΓdJ\@˜5GaLBAUk@lqUU L|,l`JR'a}Aad%}V4“VNZbl+ёS!sRe JĠLiaՂu[`50cP|vfʭ^k$b/P2;Ⴥi3 Zp3PK)ֆ y=0qP{-E3@â 85YF;ڐOXZ {+~˯KVha[I^J2_i|TrB+EfhB mSȶDNO>1w;NOҨ frR.\"K$s%r3ͽ,2ȼ.iܩcu\o2t22\Zo띓Ӵx%,;| 3zSPե‹ X kыhH&N{ϘKd g2atJ-MzalCo`p`/@m%w!A]ua?;20wuQV7W'%:r>?t+[w\r==7SSeԼw} (Efo??r܌qcdzwyܻ+zqU1Q}:zn\T]=N i?ph6oG?+D:%xKp6)k /Vw=7Rai @§W\qvսe C7*9@6!fU Ym@ltXr+-edȶd鷻7ݮB,d󥣇1j]cߓ8!5M=虗->qA;VA<j".9hw jfaO*^;BkQ ǙC$Rn0u(D ˉkp6Bր!=: !&Dj91/!Z[1.4hu.#$/2}y}&ӎ'zBĴE45}fru3#ϟGΣQ$1h&1:m*ݝMt3z(\\C.r2sZYF/ B4kqѾ p>gg0d :5 >5"c yɒRLj'GV'e %;%Դ˶3RgTru qR F8Ә=g7ҪRh>$y`]gҠ4f)c- (T uę).XNֆRz$2+`Ln `GZ!9HL5lA$F/@gc>>\Nt^3ǁsdwG-nXCzOTrO$VaG!ծsXmY鲁=vgCՃrh9V|.bR4|!0"UKtvȁX]8s,r4Zqcwƕ)-rDX~(0mf(Qj:akIxbyPȥWR0 hwBYʄ !z2:ǍO,r; `ѪPcNAe""E/1Z*@ x&;.D^!=DBC{lG,Z#3Ztpoy& &"q!['$ ӾRԞˬ-S.=hP >FF9m`|5ԶlF8 G.%+K?]b|.n @C {PНD {i1/?\lkn&z:᛫$QjӎJ~n}{q8ݼg7zw7"q[=DMki靓p2V6LQkľ]pjfݽ.dg%>(1M,]RZ]T@X=9[4`  M_*=sYòj..Dq Ƨ4uzd㖎.ʗ6lBF|NwDmph1ɯ_'lHkJWq_P;br" WIWQ"Z3mt#R[.>u*Żx}S0}Nt%ʂlWIu= IO5N?PZr~:&sYXjРL^\9u+o9%PMSk x0,\_OQ# ʤz|Nܽ|OR&zNR[=){RXIug#J$F'g$ܔSv M/y8!1xЋsЉGoO#dH{rCEf6bM9GԚttOdVLB뭮=`e_딝<g&X.ziR%R0#eH SIdYHakfJiPrB57\Qgs14k*ƸQ+q_l  bi"aj-D))uJGD0X8IbZjٲ$EkQ.V%fJj^flv՘MX0dGӓ,EZUY}Gn3:kc*DW\8%YUKU~W]_9ҁąӳ\Z=y˞ǣWOsx4!t2:~ cCΫ'>:3w>Ywۖ͸N .s// ٪V̶sA[Q=D( 6ϊ\QarB*(o3 R,EWHmfo3IUe-[|WQ>H7o?+Dʘ9XCFm'~-zaG'ػiCB27"V!T!Dޠ<ecv~0lB${6=jw5s>GyluYlq`%UrniAX1RdȀp(ߘV 5;h-n9.{/ j;I!Yhk ~1NJ]suګشtR  ߣ[8)"pRDईI[I' `L$+rR18}9 G;y99dDNNd7$ 7Ӥx4'/cFe LepiOZb sgFiz@oIP_h^`ᅥ`YjxX^J*KXP2>D:eƠdaL-_+FYV*-QU6[%4gǹ2( Bmu|Hrxkfke#&a؈u%B6pDs f^HzD4tv{ }6fs\/:Jt UWh^y# a՞1>}0[[Ip;Up\~ ( PH~YHpҀ|lPC=_Fnw*F;@H[[ʂkD0K̓Q _35Z0%j;Ǥ0>֍dLfY'ФSY4ZA+LԊo 99?LPgȂu3xv xuoeglZ),/x4x&w im|S4&GA1ZasPTY+N|)-)1c삏ҁ >J?\- 2aVGgq(/Vq'GQj3R|l<c!G>?=jy< 9aP}Y[ Bi]Z^/9u*,t[-;ԽT$lTkɌ+g%9 1#>0ЬJh㼒*pKi0XR3 +Շ9Y,T^"a3&[+;4@Fg&5#it/`锍t\ m{gJM_ ϤOKhա 5T,(eg{HGaLi2DMvѵkĢ&T8q<(k h{F+3*(X:nQ!Ggk"G$qL9aLR6J 4R^l=]hCOaQ5m6h`YU ι L-MJU纬eq%EVR.oRXh%۩5+wIB {nZ dn[5ر-a <6Jb%-̒hXpo9F<6_'')L0E#qv+ F(dku/Ix ݿtbAio[$]lR&#fclirVpMd@GpN||M;?r+ґ0!DǕAD2=ALBEy}.6ЌP,eÎ R HY H95ydRCG_gWH6Qʹ1tcOrt43pZ14֎w=#C#PHעhopd;$KAAWK>9>}1g@@QRGɯVU^׉˂AO;8A╟IUɝ]F'NV˦uxf/&Vc=ZyKcբJB4sΚ wVG dO],b+w}<)gc2ysw}$xriê,9]BԷ?s$5_@BѱuEm(#U1~G`Xf֓d)@iN {js(J1eꭤ-"Qvy}fGAn> eU,co۶;6Bh<I ([ =Rq Kʢ폀>uf#'`@pۗ ۤEh_x엗>C>Ͷw>)ޡU)@25Fk匐Ǘ>fC+0gg;f!xTa?BoLɘV8 !VA6-U+r\hT Ӂ8>zE11ĝlNYuBT>444-roxѠ ޹֑Mfײѱ l{ǗT;t a´tLixDnz>YY~K>yy6ښͯ-MaFKogRs81@x!v42*'dt(Mv]O[v]ΠuC/'hAӞ1-Mg"4 M`O (GئD&nQ Ir;Mr;Mr;Mr;m[^ & ^2m&SG k lLǎ FykkwlTvV7$2{GiY7rsGLEbBjBpiU#BZDP#5 olC $TdEz+Cw 4fɀ MU\FkeJASS})z=p}V`"ŗ4#Ef{>6Rm#eD*)HL0]ȽH*!jF0:E# P-tԃ,y띖s 'Nj!]V.92#vL,RrUݹ!Ihm;ȡHTSȐwC`HE[2C:4庫~wrzY Ds+ZAԦd R\$p/?TzEzzW.~~}d~l]Hkש=(o/rEO2ɗYfϿibŋ7 r/wo//9 mR^盷X'II;6)t\dmSog2}dd>C\,=7oY>2`GiN=S 1:QձFJWl9\ŝo~Oq~MH< f h60X Sd>׷V1hIKLWUD|_vi{8 } *^3 P|ć߾dzs]zEt)k99\=D{>hSZ>&xf. &L=\NUL6m&`#[-Xo|o*֊p# IMkj``! N eu<;T^Gؼ(3;c틗|TXњּ )}cj|5.sNiYLeܮ!p'5O[gHT|ܩpwńa/$]}PJpA)XǛ T-QnS϶*BzKR)1IӛKS*?~,sd6ef:vh$|uj*r*'Yxa]1m[g9?{v{ Ux.k?o}8G^}+qmK8b3P,G\{F}8ky|tOrwa }|KY美uZ;>TqTzL >eZAxv E)ȳ8(Trv {@K#mpg?p8P-r;;}"#=)AfDf!`V*]?`wsc k9R8=DZzN"7Qqz|Ergiˉ0pnִ+vsd5bqǃSޣV8:YiHdժT텋?WÙ#^͊]ٽ] T9ikF5j ^&HL@P qJ,|Nfz˨P"~Ks"{7R*T@VGt &U uθirl gFL}i=΂X}B)T9 EsPDTQjlA/RۉӧQ_#IgKAw:YR}lRW;7/A)TK L(16?8.`dॴV*!lN!5]3 #x]LW]bprKl!eҿfO}!̗ϲIElϳ=8ʮo&JuloCUQfe*SfoTυPҩ1a-GIy!/hُl&vpw5G%+f26_訧SM0GH?;; ]6^4(>6bF5-n}ܺ*eҚu(R#Ԗ O\O4e\LNL>~+QI,#JFSI63%߆8O\߷?̙[c'KiP&d Z\5N~׮*,WeOľYf\'و@C$fKe2NRQN AY%xF)& oRqxazcʹ߿F}- Q+ n\v0,M6РfHMY%ІЇP` E$VrBz]vakVօ y1aNoX\MқpO*$fƅjO.$.0b2e|~vB|}[?go`"3%z1Fgi6w'βe߻Pb~y_Λ0YU؋I$oqp:Z<,,Ȳ)sl.e, iHBNX] c ң=.7/'cȌz{BJ-<MXJzAŗVj@VhP]< W|XFJX.)%Pwy?j鹻I~3 %RCF0XHT RC h *pŬFP XB4R_t?U˶%tE=_Ųk;oW sԄt{V=ni?Nfo'MF+F9`KET&[ 4+ 1JZyxh(f Z):sWOqk WhDFr4E3 ,Ԉ gEa+ Ϭ \Pv} PI`2H<{C8EԦWN:#zLirA*+.qnaax(ܸ?qRs8l 5wrA.]FwY %s W-P<`"j.85Z*ɭa;G}DŽHg诳%mq_XS3VOv20MccѮ{; $e!MY .*cU;f%!+%;=unz?ݯcMuIbVHP.%4oJ>=u&1TPq bVq}lD_lHM79[0(lJ0!d!C,e9S:JhN:Z`{`nvV6VI&$Jf!$*g7MS{ߌ)L^|ڈp}S)ӏ$pvq )vS'@;m%!&i7Kb[ibqҪc䨟$jD+4#}[ZtaB"zL 1^&Wļn`U-AGbGh-%;QgٯYO!\X"~$: ^佖Tq1D%k]pu y V`Q2EQ7[=v\ZbpT„8ݕh h]i׈ + `;(GV G@bAn4HLP-t@"/5 LIHxE<^UWmi1$b>FX >Qش$_W% p0JygAZN5#B I Z{ĕ )HNM%頊WӪh`+5-`1M[U*̰o?@=#p\1}ߙnSV]Bq)_ޣX~:{T@%d1X-)}|~A.}x>¨$i$܅OXT|VLM:ʊl}YzwĘ-1 Ϳ[~fc%0 :ƘV%WE.J0XiATѢLw.yRSꖉ"(cE^0  AU1*t @0IBK4|(?BNTa)!\#cajc\9[{1tWyWӹn;r@{(m- =x¦:L' 1fz7)*EN`.*3Y|!\uS ESv!l @Vj++I砤AYwv*iC';:̣XГ 6A) FELI1n7Si kDxLiQ`Հ{S1`4a4[ڮ0,d6F\UN"@>:.x|)QJ] }cV288u@nBs_qK%uZsսuFX2u /0ԢLp7Y G"oC=pc`D6g@>Ǟ G ͉_ٕS+wH".cup^cK[؉Xg;}@Mҹ~[1>>F &)>sH::1= ZEΎ(t0jʿn0Xǔ r ͖kwX+GҴ?Ҥy s ^I*֝vΒ7LwӤ.`<LӖohh8a_Ze/4Ӵ K.8T4&t\jp*kTE䋈|qrJR$eeQ zk YI!ޞ0I/ṞEB"i)qX &Q`C>V)$)$h#[$2֎Vac<1lJRC $S}Z1*:5960zF[x?~ABCDUrzd~n Ej㼫NYeϿeL t1([e\BacRC5)au6ףxp!O} ډ$G]A; ԗ-_ F%JB4URUdyzT `@lMȱպ(R}LexӨz/[}^bBi5`^uۇw߷֢] /lb BFj-@qE(Y`vS Q`|2hM7C]aX&lM8'E] Jv]}8K%6Lb&ß PԢUD^K E%_"F.>"OQގ Pxۧ뀭ĺ۸9eS|_V\1t;P8&r|Rhg}\sy<3&`dz)kn_滟C`e 쾌ݗ1ؽ|FP0a*! adQ`NLiNe($aJh D$ H2 u]chTW4,{?7Y[a,Ѐ}Ceϓlk!uG[KDJ8( RV瑍*!eRRUDdH[AAQ(uᥣ UΓmUX v G gCpC06XӠ4'V PPH5`F1y]rQs1&P[HdaP m/d!KL7ۙJV*s1|By#NON ,j 64a0&औߔՆ&C#3Y feiVf?.>U]0v?y>~so"DUbL)6KsE᭍M姳z >\o8gc @<< ><QUΛ YoB|v$Ϊ'M#3f󥼼Nmc.׵s*0A:^Q65wˮn^Wڒ޵u+EЧJ!9|I[$h~, >cm+I;:ґ((Z>!g~3JÁ(CܖĿ]]PyYC–hF+oǜv՘Rf&TpCaB@DK.\LjZ3F4RG+3:Br,v'_+5?ve.8Kp6wf!jOԹ(4`,', Ǯ4PA5&j<oQN$EST ?\"pa37ȝ$u JtȢFVx q:'׊pIBI̸E#,ŀK CVON>?{%fJs2Üs?WxhTޡLʇZ] WONRC5eʀB|g0ђ_jdCYsn)?P%p8RlrUZX6TCe:C>PvqH3T$eZ 1 |93 G _x^@*MD>%k:n~5dwWo"f 7hb'h2o,]9nauԛȜ|X½ q,Yꎱ;kP:=CE ];i]$.TY5 ϠXF<.,*=0VJ];+~O{CdR[dsΜT]8aΜnxly KЂFJXwr[y%52_ny>R=GHy,'LrB1EP,;s:EB@Xq/jUn mq2$dZn/%H_,{Q ]@\C`*^\'- E|I(">p.)`"; ɫ&:*plHo D!i\r-b`Vh64 {2XD;H)҆2ͺ!pZ_ pXV M4&HU)РfD4% &RqI''<Ў8J擌ZEAԜ9 pKV*/³R9 2iS3TH5 @c؈b`aa/ʸV nuE;NJ!F\e-腢ZR ҅\wH:rQ-e${2 -0QY "$wW^^*BR1u@0k>ַxam"ΆLdlK/{f( klz͉~7 E0W9Mt Ɨ/u6dÔѴ ˢ/e{aT X?>6(2@*PBS10ɁE;o4Iv^x( @Id(WPyApAuZ<,EC(%UV꥗tk>2Q(R4GFG~fH /OK :X'%c4׊6w&JL?&{} θ16i)+>q`1/>dHbvtr/r"yG§]2 CȓJ:Ljoю__Zk6Kz}ӳe֬Cm? {?!n޻˻gVPc~w{yOnQAob-tν˿ųuogu<۶GE_EeQQo=G+fc78849 WMѰvC{%qyxr}nd?ffa*[5P,ӀHzL>NĴiȤY'`*E7vsD# P V$pWΕc{u+ um;]`zpsD9>.k ;֕5ǻzrxaYcLNq,{PU:ic&;{{}#9$Mt"g4P HN)vE. D!S)vi#;4yqOVvU.j-Ƨޘ*SOQ-P@)1S]S|iWQY9#'3Ķ$uA;)XJIdg,=[c kVF&;bf.F eK.C pʘI.EvM) y^ZՒʮJ9'IDiў:@[91JOR˜A*K<$ɗjPE䵮H^ 4VyH޶k3Qsů G %d1ŭo&Mpq*|) i'yirozkΩz+X4U7J l.AB8'mTPV-Yҭ ѭn>kI8hҢn]~+TwuWRcrR"*7ˠPX6MIQ&m28Z#n4+;qbBk [w^ڎܑYyWuԥ{ 1AA^I&DգEh9~(t& pCxBَ]%wZbY+d;NQunH0z˝m7~;1B9:8M -#{m53t^BRD"fC{5.6ܨcoW'p*/0)@4TqsBy xP&)F< qlꧫt;j8o6y=s<skn%z.ʺ˘nӞ|ot5k4~xMr*2d侤q޵.͍nz㛋^UjgȽE^%w/.`CjD ґ P!=UW\:lG'jBI 5Cr3¸o+ЅG\BzIPSGǏQ9@ U3ł.VVc }ӹu$4͟8N`&YfAcsi4vJN )"-Zhϲ!XJHsE-W;AC fDNd zuy۱:##8Dpa e#W{# ȲmrS F8;t(+u-c 嬧M%q2H$.J&bT/f 5!Eڜ_"_c7q!ߺL4{ oDxSϷ__rCM߰7ti#&1^wƺn! LFOoh'3f]@g r|Jާɇ(҉K;YrphJh, pzU;3.rwuCe.g8;I}m3ڮ%Ubs; be݁\%Czj\85k@IE^Z5Z4C~Ȝ_`bC@Czpt8%5D;~r@B)jP~1ZBZHD&Z-,\F`rr$t0/g_;+Qwʡd8\ kw4$Р60f>[߳]=ݬAKC#g+.oލVY.|()QAv%U`eHHp| 1%hI)ٛ-rWaR g$ӥ!Ic|K<3 b A I L2R%A5"PJ*V5)HN3_ijLcЪ4yɺcAY[J>$IQ6teL⾭Fv E5;jIt}rL2ӂ DgnqNr2 \_ibLC9 zaDeų|ֱ)U$Ќu>c55aV͚]>7K6zr=F}rT$W~3#p'|؛1fÇT&?C}72E!mJBuJ=J&q;j61K{_Jv;Qi@mO̭|"Aͼ*ͦR7a% BVFZ4J}eU>ݎ3QgUoܻllR7# BFDdTnEQło.9e,NK2bC[? _VɿʿLb<0zŔiL910*5TTS@9Tt#Rܐ&JO".F)Ic^d<3j׃'EY@*6_4[ g&t)r} .gp~~VK}wi2>_PvǷ ͗k'/J6%)'vsORO-DH;g_6mHt ]}8o^4z|]Yo#Ir+^6Tއ~ x}mIdT!OeHJf70YeFFdv`=րi$N9E;ޱl4+WJ^Wir<wa),︅BT ͽO`UuOِBs<{5(%O7usl3D%* AHLƔ뢸 &E5(0ck8-k J1纆jג ϕ c"Gc+hD|\Y C$ s^'s976i؆_>8x4y-tCCנKpQ.xV4 ժծ;A̡]üpg|ߩvҷć4}+u< ViaXr紈)dU:B3 ޷4ޗsbeoteb܋*e;FX6T~O=QAj;:Թ1JR@~Q8+&ȖO(/^0e _Ty} $/*|!m < e}䈧1 CݒS?0sA`n$BeIAI9f$ C:3JMrW#f-EjUP|Ia#xq[Dҭ/O ^좹"¨7 o +OmRg02~vX/Tsy=O_\cgH_z><Ó#(e&kZ q~uFe{?l|;!(Ez5aiZ+:Ŷ,йгi8o8!|OWdܔFUګvZJ]\^hhD jm5H0LQi6쇘I:14qr:DD>1l {b[!Syh& QT+0%TW>]Z P k->\TshMVSHɲ_ )<"HGolJ\$.cHBn [/'&#Q[Xuak7gX3J'm&*8JEͅw<1ctJc#st4Ϥ|OmaQ؋T6=fj1%f7 dRRܥn `6=~^'eS^]{Vg5Q"Mc-X$edkǂjAZO1KFK3f$a3fc< 0`Ԋ_Lp*s="7?,@4z~HٿR>{-J?z<'{;_|V`nLz_7ΟRX5/Aꧫo"z)+k&c|/,3˗8IxEwZz"܂īg|7*Y 0d%756gwp?}W*@=vJHvpnd>g؏GӦ)~Jrv@ҕB']$lE0Waxېyѳ{׸CZNeF tpi ,X4?E1>8o*e4n Vb J޲7L3dXdhPQ)n.J N]-1O:nϋ>lRitG dExLJK#n.(vX(.qv&꣡IWӡU Snj.vBC VG+5]ipJv@L4ƴ{;rWS=IpXɇ`F*H!o!>W[Y.^JClh%l.6p}4-p}B+X5!Z~. y25hWR[ iNdHXh=T&t@uHmל#ÄF;k3vcC*ڢdȽڢ\NX/֔Q^ʒ2]);qORm}t=mX*~T 04g-2_r &fDf?y_~D KKk-AT-lU> %ʅ.'G!m`L$D$< (9?ϗ7oz>)覗sM'=98숳tt.V^:kCzɬR̜ZJOۥzd8q㠪V(^(41\̉$uڇ&%»,Wd ohUTO98ExUTbqӋ*ejnQE*s2ZZQ%yAax;Wj%!z#;d >!΋m,A^ރ)5X 8QwP˾^=`Oq~-ϚZ$v5cNM^Jm*^uiYnF84o/ffKI"a"avɍ+f>C*b<ƣ/H\Z&,bm]Zesc&ޭ*,eDgt}mKCoLrT֬{΃Z" z4ԍɮRm ;=Fړ{펙_Ww?[5[La3d#G+XΎgO SuY?H5^—Bt'sKl9Gp숳`ǘI;GW軾&ty~ߡX[0J sHcX^kl^~SF8 4ZKnq<]ثBUg( ,QEVMJHc~w+pR ט'S[oGT&pshS\` 0?}{= T>u>MړK%t{s=Tj<8՟4KNkBU_rL}W7pozS?>1`?qLϓ6;CÄ0 |y^WP?]-8{SdK>qLK*L*3 [G`S|fj b3N`~Z su hְ(tEIDNiCY JB#mpBT'LP?fgF:=nNdxG[|*:YD|~ :" 2# hyQZB6Rࢢ[U wy\jl4y:D~2F;\JW&ץ6gPGσJdK)Xno$Վn@dJ׹Pt|.cJ^qef2?gD@/0dr'[uE͒s]5cR0h΃jAm ' .d  TA Whf|Z)J,Q'g4L Ҏ<ܧ-K"?yҝ}lbHu7ͤj2 PUCS<&X)ށ?=t` Ogf?'fv'ho._mߏQr0ėqZ 0=iƈFY,x8jW&'45w Y9miԫd%b_oF~ՄjVk'ڪ.W:{6MأRٸp2Vk #i$(AsȜM2KG3R-fo|3Y~_'FD[n'PH v䝢 ~K$Z*Pn+]otQ/hx_Z!83.ӛ2EcI.)'H|$.JťfQah6aR޵6r#"eإLSfAp$HمX"ٞ,o%-YwekF6"Y_Udju0{⑄ }BG&fa؁ &k* c{~h&浤_V}:1@ۻ։VƸvXI?U bEe-Jh+:4q鴐6NCJ?M[ >}2)dRɤO D "$fV:EѶY\89 i(͞)0 e"8"V+WEU]&Dvjs:Tru;Rѿ\U}NRӫ4v|M:XCg"x0j&zyQb5AoIW"QW ˊcŠL0mq\)L٥Qw=I \ 9%DQq|1OY%ˤIeI60VZ5ZD\F>e+ĝdԢg{-4pZˢe6bkɢ4J IP?*h*Q))#ɣ|I :geV@dϳ( X%,0Y{s8O㝤FNm{ <Î \ :`]ٷضxxC15H]->apG#Yߕ>J]K X+0_&[A5rTޝkNUR o٤\ۜ7Ӹ8UNQKۼ:-Y~p|W8*.W?65켨)r_FLN^"Xv+҂_OV1YKB8yډhH3j>oꏗ]/r5ixxě5 r]ϫd'Ȟ+7 Uc3\RG<3A5KiIej4f-#IֶW+4ꤛ_Q;4+FHU=XGj"S KQfIKRD),C[W'jV*?Jn 2!8<ʤ c1s׀03d01K+G-=+`""4'Kg@9e*T46<R :L?\YM-tE]mP]xȳV+Lvj|Vv.a7M K&>, huh r-5^tPgo==IVm&H>]uATX 0G2 &@F;4jH+Ro>nZ< M|s]28)AH:I=wrkSN-E,Z']zaXחry>NP\ieQ^B>uTnzXKIX[x#Sf~ vvE('8QoPKWK1Z6}Ǝe O^1DhTc4O홱Aᕚ[럮ϱBئʔlqjQb?4jt)m1lʂ|flLA@l#,W 1eYɒ`_gdPVDОP:JCWF^F}y:G ]N+kvqJ.Oy{gjٴ9zC$\IJd"֨>9'V@@ZyGx6ϧǿUF)E{n&BgiDF]9#&&G8jъV7j+D䴕/ ˛2> < bRbAf7if#:?=^}!Z0Fi7r=9£GpmD7$jHp$8/<hFD^uB$B1^lA^3;!0FiJF ݣz%1 \$b 8HĘS.99]A[aa4?̎q=aA2dh\6Q W4IS;h@pzXtIr8a Ȥ{h!=TtwcjhlH}?@X9 j4H{!]/#ߴkU%lVuY$Yg ͻI '|3n6Q! >fÃ=bDBaľH3CDpn5 \aϮmLKqixLn01hڂv(o )7ke >[vBj]UVDf-P;%kJ%zd*`!b8jd> I"#ɚHDpRqAeBlrwMWU6---ȃj~ۓ,f[`j-X\Ak)ܣiܕ Ђ]:Xr?KɡHdzR8hb4!75Օ|EŲeIi耊z≫է0&g$yND)1!Q4'SErgb{RZ$7@U\Egu ">eh9sVXuGys5kJ*8mC*:6ȟvcS0lD sW+UaON|TEKB$v"4Җ^ G*<xjIha\zg@CDZ$t<)Xf)-}flb7ncUBK=GZ3xc$5$)g[X ͭYY&l"M $/YA9>蕑)f(0fBHkXi"1XY*le6?\iJWP9hx؂j*@bІp?`ړ%bOjiipP 6g4FP-IIF^⼡ j]r6;Ͼ-}BPSWl5oL4QQԵ[FI:!~|Væ9)VJn?ifⶶnqVa <%ٛ@DT~JE?_`f.ox7{ѷ6wg6rѢͯ\lQ9R-u9WbZl߼R҂#T~tW'?;#MIe+R:gU-mOLBj;;Zhz^>u5_:SdP$.!o~>P$-\-X9?:Ԁu12-c~!X.uŪK^ެ/`fĈZ wI{ƞTOo'~6s=z{RM#ID T-5(.sxJTpBH A0Q sj 2g -`zM!to @< 8 P*qhj%;y \W.0{n)Zqcj3d>V{Jr47У·P0&k ?JUząaCAM]UT|D@IbD1qpP(%.7?*}uQ9^~RjÑTJNiS"R)~p04V&41y6z{'0 a3UŖŅ;:kX݈3Hd*V $Z9P$ֲEu[]=+tn .TS6,h~"AcYuq,,R|lˋ-: da-V d0f2BWѽ%2U&q$s*W)E(&#2Lr[H]H֊m*AeTK+b7I ,Ml +D`rƒeL8[;,$ʒx(_u6[IhRvsWuF͸pXۚM[rs8~Fb&<(^ù:7]8{YUq2=?ur{{3&G a!IR90]+a׹( T$]gQ)"1ƬMz. .I#DI>9^+ D}rE}פbl۵C`a=F{uLqxH,<N8_NAdnT̩S3guլ2 peiJ K` X0JXn\ qJ,1:I%Ʉ7 3EQؗfA/27뻇3jc҈a9#U76Xie0e\c'6ME->?a1HZgf{3zzp)FG=3$Ą cơz(1Gq-,иw|N`x L)x`h0aS S? fZ@spY_z8]nV":є cAݾL+zDx%8:@M_֪La W͐ 3þۭ%h=kʘf[0&bZ 6jH>`(>as'.]NPLxgM6Uag@8yx ,\/gTu%|0Ϩ-7Y~v~^s1Г"@ ~~/,_< +?}G5sGЉ [^lnxǔY܋ٍ YÇׅy=<_Z0#p]==O+tD+6oN}[gd {s cBZWAݠ>Cwa ddX)USo=jӖΓeD>(މdxǟ2^CM\H  ;nC -sֳƥR 8R ;`Zf-3}"fdChj\huKrt ֳc_?%rFDU.^0;Ce Z"kA#| ˕CVՀJ^'>' QfJU;NZ==O8iDh0[`0)'[NzW!A"bUC{ 5+C.Lϊ\.)SˆNA,PD]!'=Q+ІnZ^ԍVVgːYt?#9;0_=̗1atK]n~֊J:aw֌(uXe}Gu*ʔ-}*nՅq)&S[7!&aby:ny_(SI*߶Ӻu!o\Et{{֍ʦ`by:nc"8BnV=[UNI~5ҹWK#4]:!e{O3#x?6var/home/core/zuul-output/logs/kubelet.log0000644000000000000000001541267115157363732017720 0ustar rootrootMar 21 00:11:26 crc systemd[1]: Starting Kubernetes Kubelet... Mar 21 00:11:27 crc kubenswrapper[5117]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 00:11:27 crc kubenswrapper[5117]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 21 00:11:27 crc kubenswrapper[5117]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 00:11:27 crc kubenswrapper[5117]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 00:11:27 crc kubenswrapper[5117]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 21 00:11:27 crc kubenswrapper[5117]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.169538 5117 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182235 5117 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182302 5117 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182312 5117 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182320 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182329 5117 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182337 5117 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182344 5117 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182352 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182359 5117 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182366 5117 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182373 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182380 5117 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182388 5117 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182395 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182403 5117 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182410 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182417 5117 feature_gate.go:328] unrecognized feature gate: Example2 Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182425 5117 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182432 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182438 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182445 5117 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182453 5117 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182463 5117 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182476 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182485 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182493 5117 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182514 5117 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182521 5117 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182528 5117 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182536 5117 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182543 5117 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182550 5117 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182557 5117 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182564 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182571 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182578 5117 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182585 5117 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182592 5117 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182599 5117 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182606 5117 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182613 5117 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182620 5117 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182627 5117 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182634 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182641 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182648 5117 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182655 5117 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182662 5117 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182669 5117 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182676 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182683 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182689 5117 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182696 5117 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182704 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182712 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182719 5117 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182726 5117 feature_gate.go:328] unrecognized feature gate: Example Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182733 5117 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182740 5117 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182749 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182756 5117 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182763 5117 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182776 5117 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182784 5117 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182791 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182800 5117 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182807 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182817 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182824 5117 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182832 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182838 5117 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182846 5117 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182852 5117 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182859 5117 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182866 5117 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182873 5117 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182880 5117 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182887 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182894 5117 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182901 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182908 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182918 5117 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182926 5117 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182933 5117 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182940 5117 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.182947 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184123 5117 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184144 5117 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184162 5117 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184170 5117 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184211 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184221 5117 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184229 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184240 5117 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184273 5117 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184282 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184290 5117 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184298 5117 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184308 5117 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184316 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184324 5117 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184338 5117 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184345 5117 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184352 5117 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184360 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184368 5117 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184376 5117 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184383 5117 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184390 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184398 5117 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184407 5117 feature_gate.go:328] unrecognized feature gate: Example Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184414 5117 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184728 5117 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184736 5117 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184749 5117 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184757 5117 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184797 5117 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184804 5117 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184812 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184820 5117 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184830 5117 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184837 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184845 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184853 5117 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184863 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184871 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184886 5117 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184894 5117 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184901 5117 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184909 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184917 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184924 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184933 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184940 5117 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184948 5117 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184955 5117 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184962 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184970 5117 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184978 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.184992 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185000 5117 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185007 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185015 5117 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185022 5117 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185029 5117 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185041 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185048 5117 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185056 5117 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185064 5117 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185071 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185079 5117 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185096 5117 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185105 5117 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185112 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185120 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185127 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185135 5117 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185144 5117 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185151 5117 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185159 5117 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185166 5117 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185173 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185181 5117 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185189 5117 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185205 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185212 5117 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185225 5117 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185233 5117 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185244 5117 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185275 5117 feature_gate.go:328] unrecognized feature gate: Example2 Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185284 5117 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.185293 5117 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.186582 5117 flags.go:64] FLAG: --address="0.0.0.0" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187342 5117 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187416 5117 flags.go:64] FLAG: --anonymous-auth="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187435 5117 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187451 5117 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187464 5117 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187481 5117 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187498 5117 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187511 5117 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187523 5117 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187536 5117 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187547 5117 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187557 5117 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187567 5117 flags.go:64] FLAG: --cgroup-root="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187576 5117 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187586 5117 flags.go:64] FLAG: --client-ca-file="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187597 5117 flags.go:64] FLAG: --cloud-config="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187607 5117 flags.go:64] FLAG: --cloud-provider="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187618 5117 flags.go:64] FLAG: --cluster-dns="[]" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187635 5117 flags.go:64] FLAG: --cluster-domain="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187645 5117 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187658 5117 flags.go:64] FLAG: --config-dir="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187669 5117 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187681 5117 flags.go:64] FLAG: --container-log-max-files="5" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187697 5117 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187710 5117 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187721 5117 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187732 5117 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187743 5117 flags.go:64] FLAG: --contention-profiling="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187754 5117 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187765 5117 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187776 5117 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187787 5117 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187806 5117 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187817 5117 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187828 5117 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187838 5117 flags.go:64] FLAG: --enable-load-reader="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187851 5117 flags.go:64] FLAG: --enable-server="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187862 5117 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187880 5117 flags.go:64] FLAG: --event-burst="100" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187893 5117 flags.go:64] FLAG: --event-qps="50" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187904 5117 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187916 5117 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187926 5117 flags.go:64] FLAG: --eviction-hard="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187938 5117 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187947 5117 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187955 5117 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187964 5117 flags.go:64] FLAG: --eviction-soft="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187972 5117 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187980 5117 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187989 5117 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.187998 5117 flags.go:64] FLAG: --experimental-mounter-path="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188006 5117 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188014 5117 flags.go:64] FLAG: --fail-swap-on="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188022 5117 flags.go:64] FLAG: --feature-gates="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188034 5117 flags.go:64] FLAG: --file-check-frequency="20s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188043 5117 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188052 5117 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188062 5117 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188072 5117 flags.go:64] FLAG: --healthz-port="10248" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188080 5117 flags.go:64] FLAG: --help="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188089 5117 flags.go:64] FLAG: --hostname-override="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188097 5117 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188107 5117 flags.go:64] FLAG: --http-check-frequency="20s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188115 5117 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188124 5117 flags.go:64] FLAG: --image-credential-provider-config="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188133 5117 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188141 5117 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188149 5117 flags.go:64] FLAG: --image-service-endpoint="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188157 5117 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188165 5117 flags.go:64] FLAG: --kube-api-burst="100" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188202 5117 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188212 5117 flags.go:64] FLAG: --kube-api-qps="50" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188221 5117 flags.go:64] FLAG: --kube-reserved="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188229 5117 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188237 5117 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188247 5117 flags.go:64] FLAG: --kubelet-cgroups="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188282 5117 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188291 5117 flags.go:64] FLAG: --lock-file="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188299 5117 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188308 5117 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188317 5117 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188331 5117 flags.go:64] FLAG: --log-json-split-stream="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188340 5117 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188348 5117 flags.go:64] FLAG: --log-text-split-stream="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188356 5117 flags.go:64] FLAG: --logging-format="text" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188364 5117 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188373 5117 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188383 5117 flags.go:64] FLAG: --manifest-url="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188393 5117 flags.go:64] FLAG: --manifest-url-header="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188411 5117 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188426 5117 flags.go:64] FLAG: --max-open-files="1000000" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188441 5117 flags.go:64] FLAG: --max-pods="110" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188455 5117 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188467 5117 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188478 5117 flags.go:64] FLAG: --memory-manager-policy="None" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188488 5117 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188500 5117 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188515 5117 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188524 5117 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhel" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188558 5117 flags.go:64] FLAG: --node-status-max-images="50" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188566 5117 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188576 5117 flags.go:64] FLAG: --oom-score-adj="-999" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188585 5117 flags.go:64] FLAG: --pod-cidr="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188595 5117 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cc2b30e70040205c2536d01ae5c850be1ed2d775cf13249e50328e5085777977" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188612 5117 flags.go:64] FLAG: --pod-manifest-path="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188623 5117 flags.go:64] FLAG: --pod-max-pids="-1" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188635 5117 flags.go:64] FLAG: --pods-per-core="0" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188646 5117 flags.go:64] FLAG: --port="10250" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188658 5117 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188668 5117 flags.go:64] FLAG: --provider-id="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188679 5117 flags.go:64] FLAG: --qos-reserved="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188688 5117 flags.go:64] FLAG: --read-only-port="10255" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188697 5117 flags.go:64] FLAG: --register-node="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188705 5117 flags.go:64] FLAG: --register-schedulable="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188714 5117 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188731 5117 flags.go:64] FLAG: --registry-burst="10" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188740 5117 flags.go:64] FLAG: --registry-qps="5" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188749 5117 flags.go:64] FLAG: --reserved-cpus="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188760 5117 flags.go:64] FLAG: --reserved-memory="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188772 5117 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188783 5117 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188794 5117 flags.go:64] FLAG: --rotate-certificates="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188805 5117 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188817 5117 flags.go:64] FLAG: --runonce="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188828 5117 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188839 5117 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188849 5117 flags.go:64] FLAG: --seccomp-default="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188859 5117 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188869 5117 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188880 5117 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188891 5117 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188902 5117 flags.go:64] FLAG: --storage-driver-password="root" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188912 5117 flags.go:64] FLAG: --storage-driver-secure="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188923 5117 flags.go:64] FLAG: --storage-driver-table="stats" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188933 5117 flags.go:64] FLAG: --storage-driver-user="root" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188942 5117 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188953 5117 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188964 5117 flags.go:64] FLAG: --system-cgroups="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188973 5117 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.188991 5117 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189001 5117 flags.go:64] FLAG: --tls-cert-file="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189012 5117 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189025 5117 flags.go:64] FLAG: --tls-min-version="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189034 5117 flags.go:64] FLAG: --tls-private-key-file="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189042 5117 flags.go:64] FLAG: --topology-manager-policy="none" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189050 5117 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189060 5117 flags.go:64] FLAG: --topology-manager-scope="container" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189069 5117 flags.go:64] FLAG: --v="2" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189082 5117 flags.go:64] FLAG: --version="false" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189093 5117 flags.go:64] FLAG: --vmodule="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189104 5117 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.189113 5117 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189464 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189481 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189495 5117 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189511 5117 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189527 5117 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189539 5117 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189550 5117 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189560 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189569 5117 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189580 5117 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189589 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189599 5117 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189608 5117 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189618 5117 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189627 5117 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189636 5117 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189645 5117 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189655 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189664 5117 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189674 5117 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189682 5117 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189692 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189701 5117 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189711 5117 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189721 5117 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189739 5117 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189750 5117 feature_gate.go:328] unrecognized feature gate: Example Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189762 5117 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189771 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189781 5117 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189790 5117 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189800 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189808 5117 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189816 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189823 5117 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189830 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189837 5117 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189846 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189858 5117 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189866 5117 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189873 5117 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189880 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189887 5117 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189894 5117 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189901 5117 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189908 5117 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189916 5117 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189923 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189930 5117 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189938 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189946 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189969 5117 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189976 5117 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189984 5117 feature_gate.go:328] unrecognized feature gate: Example2 Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.189992 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190001 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190010 5117 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190024 5117 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190034 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190043 5117 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190052 5117 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190061 5117 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190070 5117 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190079 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190087 5117 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190095 5117 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190102 5117 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190109 5117 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190116 5117 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190126 5117 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190144 5117 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190153 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190163 5117 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190172 5117 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190182 5117 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190191 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190200 5117 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190207 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190214 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190221 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190229 5117 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190238 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190247 5117 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190291 5117 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190301 5117 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.190310 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.190326 5117 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.204791 5117 server.go:530] "Kubelet version" kubeletVersion="v1.33.5" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.204829 5117 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204915 5117 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204925 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204932 5117 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204938 5117 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204945 5117 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204950 5117 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204955 5117 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204962 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204967 5117 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204972 5117 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204977 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204982 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204987 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204992 5117 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.204997 5117 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205002 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205007 5117 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205012 5117 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205017 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205022 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205028 5117 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205033 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205038 5117 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205043 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205048 5117 feature_gate.go:328] unrecognized feature gate: Example Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205053 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205059 5117 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205064 5117 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205071 5117 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205078 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205084 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205091 5117 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205097 5117 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205103 5117 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205107 5117 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205113 5117 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205118 5117 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205123 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205127 5117 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205132 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205137 5117 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205142 5117 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205146 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205151 5117 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205156 5117 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205161 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205166 5117 feature_gate.go:328] unrecognized feature gate: Example2 Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205170 5117 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205175 5117 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205180 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205185 5117 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205190 5117 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205195 5117 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205200 5117 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205205 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205210 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205215 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205219 5117 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205224 5117 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205229 5117 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205234 5117 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205239 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205244 5117 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205274 5117 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205280 5117 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205286 5117 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205292 5117 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205297 5117 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205301 5117 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205308 5117 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205314 5117 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205319 5117 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205324 5117 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205329 5117 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205334 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205338 5117 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205344 5117 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205349 5117 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205353 5117 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205359 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205364 5117 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205370 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205376 5117 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205381 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205386 5117 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205392 5117 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.205403 5117 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205579 5117 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205590 5117 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205596 5117 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205601 5117 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205606 5117 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205612 5117 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205618 5117 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205623 5117 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205629 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205635 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205643 5117 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205652 5117 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205657 5117 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205663 5117 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205669 5117 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205674 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205680 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205685 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205690 5117 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205695 5117 feature_gate.go:328] unrecognized feature gate: GatewayAPI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205700 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205708 5117 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205713 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205718 5117 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205723 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205728 5117 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205733 5117 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205738 5117 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205743 5117 feature_gate.go:328] unrecognized feature gate: Example2 Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205748 5117 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205753 5117 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205758 5117 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205763 5117 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205768 5117 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205772 5117 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205777 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205782 5117 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205787 5117 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205792 5117 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205797 5117 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205804 5117 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205809 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205814 5117 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205823 5117 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205828 5117 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205833 5117 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205839 5117 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205844 5117 feature_gate.go:328] unrecognized feature gate: OVNObservability Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205849 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205854 5117 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205858 5117 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205864 5117 feature_gate.go:328] unrecognized feature gate: DualReplica Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205869 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205874 5117 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205879 5117 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205884 5117 feature_gate.go:328] unrecognized feature gate: SignatureStores Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205889 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205895 5117 feature_gate.go:328] unrecognized feature gate: NewOLM Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205899 5117 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205904 5117 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205910 5117 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205915 5117 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205920 5117 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205925 5117 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205930 5117 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205936 5117 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205941 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205946 5117 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205951 5117 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205956 5117 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205960 5117 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205966 5117 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205971 5117 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205976 5117 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205981 5117 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205987 5117 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205993 5117 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.205999 5117 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.206004 5117 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.206009 5117 feature_gate.go:328] unrecognized feature gate: PinnedImages Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.206014 5117 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.206019 5117 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.206025 5117 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.206030 5117 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.206035 5117 feature_gate.go:328] unrecognized feature gate: Example Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.206040 5117 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.206050 5117 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.206779 5117 server.go:962] "Client rotation is on, will bootstrap in background" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.210647 5117 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2025-12-03 08:27:53 +0000 UTC" logger="UnhandledError" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.216384 5117 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.216602 5117 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.218057 5117 server.go:1019] "Starting client certificate rotation" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.218240 5117 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.218418 5117 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.249632 5117 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.253806 5117 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.255307 5117 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.272477 5117 log.go:25] "Validated CRI v1 runtime API" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.339977 5117 log.go:25] "Validated CRI v1 image API" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.342907 5117 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.350479 5117 fs.go:135] Filesystem UUIDs: map[19e76f87-96b8-4794-9744-0b33dca22d5b:/dev/vda3 2026-03-21-00-02-37-00:/dev/sr0 5eb7c122-420e-4494-80ec-41664070d7b6:/dev/vda4 7B77-95E7:/dev/vda2] Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.350565 5117 fs.go:136] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:45 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:46 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.382212 5117 manager.go:217] Machine: {Timestamp:2026-03-21 00:11:27.378040275 +0000 UTC m=+0.671327547 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33649930240 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:80bc4fba336e4ca1bc9d28a8be52a356 SystemUUID:a07d7eeb-81b3-459f-bb75-0975d4964835 BootID:6152d979-425d-4df2-909b-43e70deffa85 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16824963072 Type:vfs Inodes:4107657 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6729986048 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:45 Capacity:3364990976 Type:vfs Inodes:821531 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6545408 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16824967168 Type:vfs Inodes:1048576 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:46 Capacity:1073741824 Type:vfs Inodes:4107657 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:9e:97:f4 Speed:0 Mtu:1500} {Name:br-int MacAddress:b2:a9:9f:57:07:84 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:9e:97:f4 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:48:b9:07 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:cc:3c:7d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ac:fe:98 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d4:2b:a6 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:e2:f7:6c:97:10:f3 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:82:d7:fe:48:05:73 Speed:0 Mtu:1500} {Name:tap0 MacAddress:5a:94:ef:e4:0c:ee Speed:10 Mtu:1500}] Topology:[{Id:0 Memory:33649930240 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.382737 5117 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.382983 5117 manager.go:233] Version: {KernelVersion:5.14.0-570.57.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20251021-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.385728 5117 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.385829 5117 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.386375 5117 topology_manager.go:138] "Creating topology manager with none policy" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.386405 5117 container_manager_linux.go:306] "Creating device plugin manager" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.386470 5117 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.388692 5117 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.389874 5117 state_mem.go:36] "Initialized new in-memory state store" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.390236 5117 server.go:1267] "Using root directory" path="/var/lib/kubelet" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.394012 5117 kubelet.go:491] "Attempting to sync node with API server" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.394069 5117 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.394091 5117 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.394108 5117 kubelet.go:397] "Adding apiserver pod source" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.394133 5117 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.397315 5117 state_checkpoint.go:81] "State checkpoint: restored pod resource state from checkpoint" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.397339 5117 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.399015 5117 state_checkpoint.go:81] "State checkpoint: restored pod resource state from checkpoint" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.399044 5117 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.400572 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.400624 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.406003 5117 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.5-3.rhaos4.20.gitd0ea985.el9" apiVersion="v1" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.406516 5117 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-server-current.pem" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.408343 5117 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409559 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409622 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409644 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409662 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409679 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409707 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409726 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409742 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409771 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409814 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.409860 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.410431 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.411496 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.411533 5117 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.412678 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.467164 5117 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.467300 5117 server.go:1295] "Started kubelet" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.467617 5117 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.467905 5117 server_v1.go:47] "podresources" method="list" useActivePods=true Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.467660 5117 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.468664 5117 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 21 00:11:27 crc systemd[1]: Started Kubernetes Kubelet. Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.471074 5117 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.474710 5117 server.go:317] "Adding debug handlers to kubelet server" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.475052 5117 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.476097 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="200ms" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.476483 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.476968 5117 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.476992 5117 volume_manager.go:295] "The desired_state_of_world populator starts" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.477018 5117 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.477699 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.478449 5117 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.243:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189eb2cad59818d6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.46720687 +0000 UTC m=+0.760494082,LastTimestamp:2026-03-21 00:11:27.46720687 +0000 UTC m=+0.760494082,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.487856 5117 factory.go:55] Registering systemd factory Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.488111 5117 factory.go:223] Registration of the systemd container factory successfully Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.488879 5117 factory.go:153] Registering CRI-O factory Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.488959 5117 factory.go:223] Registration of the crio container factory successfully Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.489140 5117 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.489184 5117 factory.go:103] Registering Raw factory Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.489236 5117 manager.go:1196] Started watching for new ooms in manager Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.493428 5117 manager.go:319] Starting recovery of all containers Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.497245 5117 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.548826 5117 manager.go:324] Recovery completed Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.559437 5117 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/system.slice/crc-routes-controller.service": inotify_add_watch /sys/fs/cgroup/system.slice/crc-routes-controller.service: no such file or directory Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.559543 5117 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/system.slice/crc-wait-node-ready.service": inotify_add_watch /sys/fs/cgroup/system.slice/crc-wait-node-ready.service: no such file or directory Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.560537 5117 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/system.slice/ocp-clusterid.service": inotify_add_watch /sys/fs/cgroup/system.slice/ocp-clusterid.service: no such file or directory Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.560595 5117 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/system.slice/ocp-mco-sshkey.service": inotify_add_watch /sys/fs/cgroup/system.slice/ocp-mco-sshkey.service: no such file or directory Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.560636 5117 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/system.slice/ocp-userpasswords.service": inotify_add_watch /sys/fs/cgroup/system.slice/ocp-userpasswords.service: no such file or directory Mar 21 00:11:27 crc kubenswrapper[5117]: W0321 00:11:27.567161 5117 watcher.go:93] Error while processing event ("/sys/fs/cgroup/system.slice/ocp-wait-apiservices-available.service": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/system.slice/ocp-wait-apiservices-available.service: no such file or directory Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.577244 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.577666 5117 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.577730 5117 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.577774 5117 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.577789 5117 kubelet.go:2451] "Starting kubelet main sync loop" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.577912 5117 kubelet.go:2475] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.579578 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.586288 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.588234 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.588504 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.588527 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.590148 5117 cpu_manager.go:222] "Starting CPU manager" policy="none" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.590168 5117 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.590193 5117 state_mem.go:36] "Initialized new in-memory state store" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.594966 5117 policy_none.go:49] "None policy: Start" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.595021 5117 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.595051 5117 state_mem.go:35] "Initializing new in-memory state store" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605004 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="584e1f4a-8205-47d7-8efb-3afc6017c4c9" volumeName="kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605073 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605090 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a52afe44-fb37-46ed-a1f8-bf39727a3cbe" volumeName="kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605106 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605119 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605132 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605145 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605158 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605173 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605186 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605198 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f0bc7fcb0822a2c13eb2d22cd8c0641" volumeName="kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-ca-trust-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605209 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605220 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605232 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605268 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605283 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" volumeName="kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605295 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605309 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d7e8f42f-dc0e-424b-bb56-5ec849834888" volumeName="kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605323 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605337 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605351 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605365 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ee8fbd3-1f81-4666-96da-5afc70819f1a" volumeName="kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605377 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605389 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605402 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605414 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605427 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605442 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605463 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3a14caf222afb62aaabdc47808b6f944" volumeName="kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605476 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="593a3561-7760-45c5-8f91-5aaef7475d0f" volumeName="kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605500 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20c5c5b4bed930554494851fe3cb2b2a" volumeName="kubernetes.io/empty-dir/20c5c5b4bed930554494851fe3cb2b2a-tmp-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605516 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605544 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605558 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605571 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605586 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605597 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605609 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605625 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="149b3c48-e17c-4a66-a835-d86dabf6ff13" volumeName="kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605639 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="16bdd140-dce1-464c-ab47-dd5798d1d256" volumeName="kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605651 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605665 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="17b87002-b798-480a-8e17-83053d698239" volumeName="kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605680 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31fa8943-81cc-4750-a0b7-0fa9ab5af883" volumeName="kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605695 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605709 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" volumeName="kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605730 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cc85e424-18b2-4924-920b-bd291a8c4b01" volumeName="kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605743 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605756 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d7e8f42f-dc0e-424b-bb56-5ec849834888" volumeName="kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605769 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7e2c886-118e-43bb-bef1-c78134de392b" volumeName="kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605806 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605821 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605837 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="593a3561-7760-45c5-8f91-5aaef7475d0f" volumeName="kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605851 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7599e0b6-bddf-4def-b7f2-0b32206e8651" volumeName="kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605891 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605904 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ce090a97-9ab6-4c40-a719-64ff2acd9778" volumeName="kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605919 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="428b39f5-eb1c-4f65-b7a4-eeb6e84860cc" volumeName="kubernetes.io/configmap/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-iptables-alerter-script" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605942 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="42a11a02-47e1-488f-b270-2679d3298b0e" volumeName="kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605955 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605968 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605983 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.605995 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606008 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606021 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606034 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606046 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606061 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606074 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b605f283-6f2e-42da-a838-54421690f7d0" volumeName="kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606089 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a208c9c2-333b-4b4a-be0d-bc32ec38a821" volumeName="kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606101 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b4750666-1362-4001-abd0-6f89964cc621" volumeName="kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606113 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606125 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606172 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-ovnkube-identity-cm" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606188 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="16bdd140-dce1-464c-ab47-dd5798d1d256" volumeName="kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606203 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606219 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606234 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b4750666-1362-4001-abd0-6f89964cc621" volumeName="kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606246 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" volumeName="kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606282 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/projected/fc4541ce-7789-4670-bc75-5c2868e52ce0-kube-api-access-8nt2j" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606296 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606310 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606323 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606337 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f863fff9-286a-45fa-b8f0-8a86994b8440" volumeName="kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606351 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af41de71-79cf-4590-bbe9-9e8b848862cb" volumeName="kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606365 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606378 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606390 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" volumeName="kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606405 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09cfa50b-4138-4585-a53e-64dd3ab73335" volumeName="kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606421 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606435 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606449 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" volumeName="kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606462 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606477 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606489 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f0bc7fcb0822a2c13eb2d22cd8c0641" volumeName="kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-var-run-kubernetes" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606503 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606517 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606532 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7599e0b6-bddf-4def-b7f2-0b32206e8651" volumeName="kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606546 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606560 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" volumeName="kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606574 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606587 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606601 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="584e1f4a-8205-47d7-8efb-3afc6017c4c9" volumeName="kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606614 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7599e0b6-bddf-4def-b7f2-0b32206e8651" volumeName="kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.606627 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610600 5117 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b1264ac67579ad07e7e9003054d44fe40dd55285a4b2f7dc74e48be1aee0868a/globalmount" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610638 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610654 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610669 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0dd0fbac-8c0d-4228-8faa-abbeedabf7db" volumeName="kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610685 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610700 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610714 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="81e39f7b-62e4-4fc9-992a-6535ce127a02" volumeName="kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610727 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a208c9c2-333b-4b4a-be0d-bc32ec38a821" volumeName="kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610740 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610754 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610785 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610799 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610816 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0effdbcf-dd7d-404d-9d48-77536d665a5d" volumeName="kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610831 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610845 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610857 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610871 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610887 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" volumeName="kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610902 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610917 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="428b39f5-eb1c-4f65-b7a4-eeb6e84860cc" volumeName="kubernetes.io/projected/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-kube-api-access-dsgwk" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610932 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610944 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610958 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cc85e424-18b2-4924-920b-bd291a8c4b01" volumeName="kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610973 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6077b63e-53a2-4f96-9d56-1ce0324e4913" volumeName="kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610986 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.610998 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611012 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b605f283-6f2e-42da-a838-54421690f7d0" volumeName="kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611026 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611041 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611054 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="593a3561-7760-45c5-8f91-5aaef7475d0f" volumeName="kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611067 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5ebfebf6-3ecd-458e-943f-bb25b52e2718" volumeName="kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611080 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611095 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="16bdd140-dce1-464c-ab47-dd5798d1d256" volumeName="kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611110 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611124 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611138 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611151 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="584e1f4a-8205-47d7-8efb-3afc6017c4c9" volumeName="kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611164 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611178 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cc85e424-18b2-4924-920b-bd291a8c4b01" volumeName="kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611191 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611203 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611217 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611231 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611246 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611283 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611296 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b4750666-1362-4001-abd0-6f89964cc621" volumeName="kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611310 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611324 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01080b46-74f1-4191-8755-5152a57b3b25" volumeName="kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611341 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611356 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="94a6e063-3d1a-4d44-875d-185291448c31" volumeName="kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611372 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b638b8f4bb0070e40528db779baf6a2" volumeName="kubernetes.io/empty-dir/0b638b8f4bb0070e40528db779baf6a2-tmp" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611391 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611410 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="34177974-8d82-49d2-a763-391d0df3bbd8" volumeName="kubernetes.io/projected/34177974-8d82-49d2-a763-391d0df3bbd8-kube-api-access-m7xz2" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611427 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f0bc7fcb0822a2c13eb2d22cd8c0641" volumeName="kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-tmp-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611441 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611457 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="149b3c48-e17c-4a66-a835-d86dabf6ff13" volumeName="kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611471 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611485 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611507 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611521 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611536 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0dd0fbac-8c0d-4228-8faa-abbeedabf7db" volumeName="kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611550 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611567 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e093be35-bb62-4843-b2e8-094545761610" volumeName="kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611581 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611593 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6077b63e-53a2-4f96-9d56-1ce0324e4913" volumeName="kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611607 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611620 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611632 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611647 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" volumeName="kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611660 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611689 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" volumeName="kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611702 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611714 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611727 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611740 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611753 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" volumeName="kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611772 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611785 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b605f283-6f2e-42da-a838-54421690f7d0" volumeName="kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611798 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611812 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611832 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611845 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01080b46-74f1-4191-8755-5152a57b3b25" volumeName="kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611857 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611869 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611886 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09cfa50b-4138-4585-a53e-64dd3ab73335" volumeName="kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611901 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31fa8943-81cc-4750-a0b7-0fa9ab5af883" volumeName="kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611915 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="42a11a02-47e1-488f-b270-2679d3298b0e" volumeName="kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611930 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611968 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ce090a97-9ab6-4c40-a719-64ff2acd9778" volumeName="kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.611984 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" volumeName="kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612000 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3a14caf222afb62aaabdc47808b6f944" volumeName="kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612015 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612031 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612046 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d7e8f42f-dc0e-424b-bb56-5ec849834888" volumeName="kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612059 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="34177974-8d82-49d2-a763-391d0df3bbd8" volumeName="kubernetes.io/secret/34177974-8d82-49d2-a763-391d0df3bbd8-metrics-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612073 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612087 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612099 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612114 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612127 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612139 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a52afe44-fb37-46ed-a1f8-bf39727a3cbe" volumeName="kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612153 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5ebfebf6-3ecd-458e-943f-bb25b52e2718" volumeName="kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612166 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612181 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612195 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31fa8943-81cc-4750-a0b7-0fa9ab5af883" volumeName="kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612219 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612231 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612244 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" volumeName="kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612278 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612292 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612306 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ee8fbd3-1f81-4666-96da-5afc70819f1a" volumeName="kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612320 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="81e39f7b-62e4-4fc9-992a-6535ce127a02" volumeName="kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612333 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ce090a97-9ab6-4c40-a719-64ff2acd9778" volumeName="kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612345 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612359 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612372 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="94a6e063-3d1a-4d44-875d-185291448c31" volumeName="kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612388 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01080b46-74f1-4191-8755-5152a57b3b25" volumeName="kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612407 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09cfa50b-4138-4585-a53e-64dd3ab73335" volumeName="kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612421 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612434 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612447 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="149b3c48-e17c-4a66-a835-d86dabf6ff13" volumeName="kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612460 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612536 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" volumeName="kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612550 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612564 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612581 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612596 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612611 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612622 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="81e39f7b-62e4-4fc9-992a-6535ce127a02" volumeName="kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612634 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612646 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612657 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-env-overrides" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612669 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612681 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612693 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612706 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="94a6e063-3d1a-4d44-875d-185291448c31" volumeName="kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612717 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7e2c886-118e-43bb-bef1-c78134de392b" volumeName="kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612729 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6077b63e-53a2-4f96-9d56-1ce0324e4913" volumeName="kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612741 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612752 5117 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/secret/fc4541ce-7789-4670-bc75-5c2868e52ce0-webhook-cert" seLinuxMountContext="" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612767 5117 reconstruct.go:97] "Volume reconstruction finished" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.612776 5117 reconciler.go:26] "Reconciler: start to sync state" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.648404 5117 manager.go:341] "Starting Device Plugin manager" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.648728 5117 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.648764 5117 server.go:85] "Starting device plugin registration server" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.649475 5117 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.649516 5117 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.649733 5117 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.650049 5117 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.650079 5117 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.653773 5117 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.653858 5117 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.677875 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="400ms" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.678779 5117 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.679125 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.680151 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.680240 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.680296 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.681523 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.681862 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.682005 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.683007 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.683010 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.683124 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.683152 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.683048 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.683214 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.685191 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.685353 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.685403 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.685838 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.685878 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.685894 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.686766 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.686995 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687076 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687564 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687626 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687652 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687735 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687764 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687780 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687761 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687891 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.687952 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.689769 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.690397 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.690469 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.691309 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.691358 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.691371 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.691493 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.691578 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.691697 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.692353 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.692399 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.693608 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.693808 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.693845 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.709779 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.722983 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.744376 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.750123 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.751390 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.751437 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.751453 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.751489 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.752270 5117 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.771840 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.776429 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.814708 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815320 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0b638b8f4bb0070e40528db779baf6a2-tmp\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815353 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815382 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-resource-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815402 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-data-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815425 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-auto-backup-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-etcd-auto-backup-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815448 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815475 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-ca-trust-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815491 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-run-kubernetes\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-var-run-kubernetes\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815511 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815529 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815544 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815564 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815581 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815596 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815612 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-cert-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815629 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-log-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815648 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/20c5c5b4bed930554494851fe3cb2b2a-tmp-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815689 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815715 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815738 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-tmp-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815760 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-static-pod-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.815775 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-usr-local-bin\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.816756 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.816948 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.817229 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-ca-trust-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.817351 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0b638b8f4bb0070e40528db779baf6a2-tmp\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.817446 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-run-kubernetes\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-var-run-kubernetes\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.817451 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-tmp-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.817934 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/20c5c5b4bed930554494851fe3cb2b2a-tmp-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917131 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917304 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917364 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917420 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917420 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917470 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917574 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-cert-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917603 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-log-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917609 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917627 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917699 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917710 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917748 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917753 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-static-pod-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917782 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-cert-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917788 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-usr-local-bin\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917815 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-log-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917827 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917849 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917880 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917920 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917928 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-resource-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917957 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917975 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-data-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917995 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-usr-local-bin\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.917887 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-static-pod-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.918011 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-auto-backup-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-etcd-auto-backup-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.918045 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.918085 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-resource-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.918119 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-data-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.918157 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.918232 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-auto-backup-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-etcd-auto-backup-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.952955 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.954027 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.954080 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.954101 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:27 crc kubenswrapper[5117]: I0321 00:11:27.954141 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:11:27 crc kubenswrapper[5117]: E0321 00:11:27.954852 5117 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.011375 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.024757 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.044966 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 00:11:28 crc kubenswrapper[5117]: W0321 00:11:28.069851 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f0bc7fcb0822a2c13eb2d22cd8c0641.slice/crio-b1a6e65127799101f1d92f108350f74728c35e039c6fb54b7ba6dd02329243e6 WatchSource:0}: Error finding container b1a6e65127799101f1d92f108350f74728c35e039c6fb54b7ba6dd02329243e6: Status 404 returned error can't find the container with id b1a6e65127799101f1d92f108350f74728c35e039c6fb54b7ba6dd02329243e6 Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.073329 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 21 00:11:28 crc kubenswrapper[5117]: W0321 00:11:28.073623 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b638b8f4bb0070e40528db779baf6a2.slice/crio-cdcde59a6eaf0ca52cd7a3f6c1aa4ec2b1e7daa69ee2091adc2166978cafb23f WatchSource:0}: Error finding container cdcde59a6eaf0ca52cd7a3f6c1aa4ec2b1e7daa69ee2091adc2166978cafb23f: Status 404 returned error can't find the container with id cdcde59a6eaf0ca52cd7a3f6c1aa4ec2b1e7daa69ee2091adc2166978cafb23f Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.076817 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.077735 5117 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 00:11:28 crc kubenswrapper[5117]: E0321 00:11:28.078883 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="800ms" Mar 21 00:11:28 crc kubenswrapper[5117]: W0321 00:11:28.091755 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e08c320b1e9e2405e6e0107bdf7eeb4.slice/crio-ffeb04d8ce2e41bdbc3e44128b4d28ee97aa5fb45258bd3b7c98038e4ce924c8 WatchSource:0}: Error finding container ffeb04d8ce2e41bdbc3e44128b4d28ee97aa5fb45258bd3b7c98038e4ce924c8: Status 404 returned error can't find the container with id ffeb04d8ce2e41bdbc3e44128b4d28ee97aa5fb45258bd3b7c98038e4ce924c8 Mar 21 00:11:28 crc kubenswrapper[5117]: W0321 00:11:28.099365 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20c5c5b4bed930554494851fe3cb2b2a.slice/crio-c5d0ceb1f3df5f4b9ef6f7a87d8e895fa17f890dfb45f5e889c6cc0588ab366e WatchSource:0}: Error finding container c5d0ceb1f3df5f4b9ef6f7a87d8e895fa17f890dfb45f5e889c6cc0588ab366e: Status 404 returned error can't find the container with id c5d0ceb1f3df5f4b9ef6f7a87d8e895fa17f890dfb45f5e889c6cc0588ab366e Mar 21 00:11:28 crc kubenswrapper[5117]: W0321 00:11:28.100968 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a14caf222afb62aaabdc47808b6f944.slice/crio-e2b41ef2b7ec2b2e871fd98999eefd35c78b780c75a53b8a9f95fa3758b5f4af WatchSource:0}: Error finding container e2b41ef2b7ec2b2e871fd98999eefd35c78b780c75a53b8a9f95fa3758b5f4af: Status 404 returned error can't find the container with id e2b41ef2b7ec2b2e871fd98999eefd35c78b780c75a53b8a9f95fa3758b5f4af Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.355681 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.357035 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.357082 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.357097 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.357132 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:11:28 crc kubenswrapper[5117]: E0321 00:11:28.357826 5117 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.414406 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.584110 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"b1a6e65127799101f1d92f108350f74728c35e039c6fb54b7ba6dd02329243e6"} Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.585377 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"e2b41ef2b7ec2b2e871fd98999eefd35c78b780c75a53b8a9f95fa3758b5f4af"} Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.586549 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"c5d0ceb1f3df5f4b9ef6f7a87d8e895fa17f890dfb45f5e889c6cc0588ab366e"} Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.588015 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"4e08c320b1e9e2405e6e0107bdf7eeb4","Type":"ContainerStarted","Data":"ffeb04d8ce2e41bdbc3e44128b4d28ee97aa5fb45258bd3b7c98038e4ce924c8"} Mar 21 00:11:28 crc kubenswrapper[5117]: I0321 00:11:28.588942 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"cdcde59a6eaf0ca52cd7a3f6c1aa4ec2b1e7daa69ee2091adc2166978cafb23f"} Mar 21 00:11:28 crc kubenswrapper[5117]: E0321 00:11:28.660008 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 21 00:11:28 crc kubenswrapper[5117]: E0321 00:11:28.845702 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 21 00:11:28 crc kubenswrapper[5117]: E0321 00:11:28.853900 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 21 00:11:28 crc kubenswrapper[5117]: E0321 00:11:28.880187 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="1.6s" Mar 21 00:11:28 crc kubenswrapper[5117]: E0321 00:11:28.982020 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.158041 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.160165 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.160243 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.160302 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.160361 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:11:29 crc kubenswrapper[5117]: E0321 00:11:29.161117 5117 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.267099 5117 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Mar 21 00:11:29 crc kubenswrapper[5117]: E0321 00:11:29.268579 5117 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.414317 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.593395 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"94684ef43a4caf326c026a0b8b43fe90d3f86c9c60562f24fa305fcfc6f70788"} Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.593636 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"ad24a3d986e06d5c3f1905aa7e44d8fcc888cb7acc101ca3cbb39adac8684500"} Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.595433 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a" exitCode=0 Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.595573 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a"} Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.595638 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.596276 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.596310 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.596321 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:29 crc kubenswrapper[5117]: E0321 00:11:29.596516 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.598866 5117 generic.go:358] "Generic (PLEG): container finished" podID="20c5c5b4bed930554494851fe3cb2b2a" containerID="1f7effd8d591e2ac11bcdb21df51a7b48ab90a589f431a721326bea14181aa6e" exitCode=0 Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.598973 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerDied","Data":"1f7effd8d591e2ac11bcdb21df51a7b48ab90a589f431a721326bea14181aa6e"} Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.599105 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.599276 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.599684 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.599717 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.599727 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:29 crc kubenswrapper[5117]: E0321 00:11:29.599917 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.600014 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.600069 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.600086 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.603278 5117 generic.go:358] "Generic (PLEG): container finished" podID="4e08c320b1e9e2405e6e0107bdf7eeb4" containerID="000182f0824f8c8f7487c695f547ec126efb5f50836571807732678886484fe4" exitCode=0 Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.603381 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"4e08c320b1e9e2405e6e0107bdf7eeb4","Type":"ContainerDied","Data":"000182f0824f8c8f7487c695f547ec126efb5f50836571807732678886484fe4"} Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.603540 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.604703 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.604728 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.604741 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:29 crc kubenswrapper[5117]: E0321 00:11:29.604918 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.606051 5117 generic.go:358] "Generic (PLEG): container finished" podID="0b638b8f4bb0070e40528db779baf6a2" containerID="c6f7650b6653e951c6fe6e6883d205ee80691d8683967540c1766ef0d1df7f36" exitCode=0 Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.606092 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerDied","Data":"c6f7650b6653e951c6fe6e6883d205ee80691d8683967540c1766ef0d1df7f36"} Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.606288 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.606858 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.606902 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:29 crc kubenswrapper[5117]: I0321 00:11:29.606917 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:29 crc kubenswrapper[5117]: E0321 00:11:29.607082 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:30 crc kubenswrapper[5117]: E0321 00:11:30.111129 5117 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.243:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189eb2cad59818d6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.46720687 +0000 UTC m=+0.760494082,LastTimestamp:2026-03-21 00:11:27.46720687 +0000 UTC m=+0.760494082,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.413818 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.243:6443: connect: connection refused Mar 21 00:11:30 crc kubenswrapper[5117]: E0321 00:11:30.481402 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="3.2s" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.610640 5117 generic.go:358] "Generic (PLEG): container finished" podID="20c5c5b4bed930554494851fe3cb2b2a" containerID="c2f608613c2ee33e9d147e0481d3f47512c4db2c0e5fe416c84ad073975f3d5f" exitCode=0 Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.610786 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerDied","Data":"c2f608613c2ee33e9d147e0481d3f47512c4db2c0e5fe416c84ad073975f3d5f"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.610898 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.612450 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"4e08c320b1e9e2405e6e0107bdf7eeb4","Type":"ContainerStarted","Data":"c2514a311c8df4b1faca8a81fa4c3becfae5f7e05e32fe441cc7d61e4b3fbf2c"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.612529 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.612565 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.612576 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.612594 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:30 crc kubenswrapper[5117]: E0321 00:11:30.612842 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.613962 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.614002 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.614013 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:30 crc kubenswrapper[5117]: E0321 00:11:30.614271 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.621489 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"cb3b6dc02877e4af1eb1e9c2de8b9041ba203d71b6e52185287a6cd4c3ba28b2"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.621549 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"3a5498baab91a2da6643fcdc8b35141226196ffe302f6c2f38a266e518799587"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.621566 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"af1c6802eebfabdf424eae7ad9eecfcaaffcd44a0d2cfad7300f2ff7a532cadf"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.622303 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.629151 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.629207 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.629225 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:30 crc kubenswrapper[5117]: E0321 00:11:30.629551 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.632670 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"ad515cb5c17c39877403cc2a002a95cf96b2d32f46c7cb4b753d390ad268c3dd"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.632702 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"d9eee0d1d08899481bf9b5b0fe5429a679042a29b12e717e65d1a11a9ab2d79e"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.632842 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.633677 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.633734 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.633751 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:30 crc kubenswrapper[5117]: E0321 00:11:30.634113 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.635648 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.635701 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.635720 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae"} Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.635730 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd"} Mar 21 00:11:30 crc kubenswrapper[5117]: E0321 00:11:30.705561 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.762005 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.764896 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.764946 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.764968 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:30 crc kubenswrapper[5117]: I0321 00:11:30.764996 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:11:30 crc kubenswrapper[5117]: E0321 00:11:30.765997 5117 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.243:6443: connect: connection refused" node="crc" Mar 21 00:11:30 crc kubenswrapper[5117]: E0321 00:11:30.914853 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 21 00:11:31 crc kubenswrapper[5117]: E0321 00:11:31.262491 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.243:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.641272 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"d9ad19ec568430350de9f8bfeeebca98b59e90ff9d3b7cc0c580f7b4d48e26d4"} Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.641450 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.641901 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.641922 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.641931 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:31 crc kubenswrapper[5117]: E0321 00:11:31.642101 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.644200 5117 generic.go:358] "Generic (PLEG): container finished" podID="20c5c5b4bed930554494851fe3cb2b2a" containerID="d6e7a9fe3510c2f332f71cb736921b49b751c42424d2cbb5f6e9cb6c3726555e" exitCode=0 Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.644310 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.644492 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.644590 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerDied","Data":"d6e7a9fe3510c2f332f71cb736921b49b751c42424d2cbb5f6e9cb6c3726555e"} Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.644650 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.644797 5117 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.644823 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.645409 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.645430 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.645438 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:31 crc kubenswrapper[5117]: E0321 00:11:31.645602 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.645763 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.645776 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.645785 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:31 crc kubenswrapper[5117]: E0321 00:11:31.645924 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.646163 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.646183 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.646191 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:31 crc kubenswrapper[5117]: E0321 00:11:31.646416 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.646712 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.646731 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:31 crc kubenswrapper[5117]: I0321 00:11:31.646738 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:31 crc kubenswrapper[5117]: E0321 00:11:31.646907 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.007671 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.470780 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652148 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"ed17e61de6c00e7e405bc7beaa0ce1f7e26b6a4ec68e83ebee1f4c7ec0d3f27c"} Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652221 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"825d8b369cbcdc65a64926c868202d0b3b9253162f40994fa0f447a5beeede1b"} Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652241 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"f6e2280de7b69beabb5b164fe42e7440579b41b13e07f84fcd7a6ad773218de2"} Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652290 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"b44d38825bd92498e91533afc0db62985336c2b9293ccdc5221262dcacd0acde"} Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652299 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652403 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652439 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652864 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652920 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.652933 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:32 crc kubenswrapper[5117]: E0321 00:11:32.653405 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.653755 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.653810 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:32 crc kubenswrapper[5117]: I0321 00:11:32.653823 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:32 crc kubenswrapper[5117]: E0321 00:11:32.654204 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.661473 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"b743995a17c55f75fba00e6815f49b6fc598d27f4c4c81452bad9ca5806d9f80"} Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.661725 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.661725 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.662653 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.662704 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.662718 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.662769 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.662823 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.662845 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:33 crc kubenswrapper[5117]: E0321 00:11:33.663029 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:33 crc kubenswrapper[5117]: E0321 00:11:33.663644 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.665748 5117 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.966526 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.967904 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.967970 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.967991 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:33 crc kubenswrapper[5117]: I0321 00:11:33.968046 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:11:34 crc kubenswrapper[5117]: I0321 00:11:34.543883 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-etcd/etcd-crc" Mar 21 00:11:34 crc kubenswrapper[5117]: I0321 00:11:34.665788 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:34 crc kubenswrapper[5117]: I0321 00:11:34.666614 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:34 crc kubenswrapper[5117]: I0321 00:11:34.666645 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:34 crc kubenswrapper[5117]: I0321 00:11:34.666657 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:34 crc kubenswrapper[5117]: E0321 00:11:34.667097 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.044859 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.045214 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.046395 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.046431 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.046449 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:35 crc kubenswrapper[5117]: E0321 00:11:35.046938 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.463350 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.463601 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.464714 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.464763 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.464780 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:35 crc kubenswrapper[5117]: E0321 00:11:35.465131 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.667392 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.668300 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.668348 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.668369 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:35 crc kubenswrapper[5117]: E0321 00:11:35.669056 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.915446 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.915972 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.917025 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.917091 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:35 crc kubenswrapper[5117]: I0321 00:11:35.917117 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:35 crc kubenswrapper[5117]: E0321 00:11:35.917726 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.203415 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.203802 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.205052 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.205122 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.205142 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:37 crc kubenswrapper[5117]: E0321 00:11:37.205729 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.600565 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.600801 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.601754 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.601817 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.601838 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:37 crc kubenswrapper[5117]: E0321 00:11:37.602404 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.608284 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:37 crc kubenswrapper[5117]: E0321 00:11:37.654094 5117 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.672978 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.673703 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.673753 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:37 crc kubenswrapper[5117]: I0321 00:11:37.673770 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:37 crc kubenswrapper[5117]: E0321 00:11:37.674188 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:38 crc kubenswrapper[5117]: I0321 00:11:38.653393 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 21 00:11:38 crc kubenswrapper[5117]: I0321 00:11:38.653999 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:38 crc kubenswrapper[5117]: I0321 00:11:38.655363 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:38 crc kubenswrapper[5117]: I0321 00:11:38.655475 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:38 crc kubenswrapper[5117]: I0321 00:11:38.655495 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:38 crc kubenswrapper[5117]: E0321 00:11:38.656562 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:38 crc kubenswrapper[5117]: I0321 00:11:38.916389 5117 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 00:11:38 crc kubenswrapper[5117]: I0321 00:11:38.916517 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 00:11:41 crc kubenswrapper[5117]: I0321 00:11:41.414577 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 21 00:11:41 crc kubenswrapper[5117]: I0321 00:11:41.960916 5117 trace.go:236] Trace[919710075]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Mar-2026 00:11:31.958) (total time: 10002ms): Mar 21 00:11:41 crc kubenswrapper[5117]: Trace[919710075]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (00:11:41.960) Mar 21 00:11:41 crc kubenswrapper[5117]: Trace[919710075]: [10.002234706s] [10.002234706s] END Mar 21 00:11:41 crc kubenswrapper[5117]: E0321 00:11:41.961007 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.008143 5117 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.008312 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.166465 5117 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.166563 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.687609 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/0.log" Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.690105 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="d9ad19ec568430350de9f8bfeeebca98b59e90ff9d3b7cc0c580f7b4d48e26d4" exitCode=255 Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.690212 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"d9ad19ec568430350de9f8bfeeebca98b59e90ff9d3b7cc0c580f7b4d48e26d4"} Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.690622 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.691522 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.691581 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.691600 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:42 crc kubenswrapper[5117]: E0321 00:11:42.692088 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:42 crc kubenswrapper[5117]: I0321 00:11:42.692638 5117 scope.go:117] "RemoveContainer" containerID="d9ad19ec568430350de9f8bfeeebca98b59e90ff9d3b7cc0c580f7b4d48e26d4" Mar 21 00:11:43 crc kubenswrapper[5117]: E0321 00:11:43.682533 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Mar 21 00:11:43 crc kubenswrapper[5117]: I0321 00:11:43.695615 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/0.log" Mar 21 00:11:43 crc kubenswrapper[5117]: I0321 00:11:43.699800 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2"} Mar 21 00:11:43 crc kubenswrapper[5117]: I0321 00:11:43.700143 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:43 crc kubenswrapper[5117]: I0321 00:11:43.701014 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:43 crc kubenswrapper[5117]: I0321 00:11:43.701101 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:43 crc kubenswrapper[5117]: I0321 00:11:43.701129 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:43 crc kubenswrapper[5117]: E0321 00:11:43.701768 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.703524 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.704106 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/0.log" Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.705822 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2" exitCode=255 Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.705887 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2"} Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.705943 5117 scope.go:117] "RemoveContainer" containerID="d9ad19ec568430350de9f8bfeeebca98b59e90ff9d3b7cc0c580f7b4d48e26d4" Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.706139 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.706744 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.706776 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.706786 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:44 crc kubenswrapper[5117]: E0321 00:11:44.707089 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:44 crc kubenswrapper[5117]: I0321 00:11:44.707366 5117 scope.go:117] "RemoveContainer" containerID="fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2" Mar 21 00:11:44 crc kubenswrapper[5117]: E0321 00:11:44.707551 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:11:45 crc kubenswrapper[5117]: I0321 00:11:45.096320 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:45 crc kubenswrapper[5117]: I0321 00:11:45.710216 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Mar 21 00:11:45 crc kubenswrapper[5117]: I0321 00:11:45.711957 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:45 crc kubenswrapper[5117]: I0321 00:11:45.712678 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:45 crc kubenswrapper[5117]: I0321 00:11:45.712719 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:45 crc kubenswrapper[5117]: I0321 00:11:45.712735 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:45 crc kubenswrapper[5117]: E0321 00:11:45.713206 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:45 crc kubenswrapper[5117]: I0321 00:11:45.713550 5117 scope.go:117] "RemoveContainer" containerID="fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2" Mar 21 00:11:45 crc kubenswrapper[5117]: E0321 00:11:45.713780 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:11:46 crc kubenswrapper[5117]: E0321 00:11:46.645067 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.017233 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.017594 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.018592 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.018668 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.018695 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.019396 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.019881 5117 scope.go:117] "RemoveContainer" containerID="fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.020323 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.022905 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.153127 5117 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.165412 5117 trace.go:236] Trace[797722025]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Mar-2026 00:11:36.269) (total time: 10895ms): Mar 21 00:11:47 crc kubenswrapper[5117]: Trace[797722025]: ---"Objects listed" error:csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope 10895ms (00:11:47.165) Mar 21 00:11:47 crc kubenswrapper[5117]: Trace[797722025]: [10.895773297s] [10.895773297s] END Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.165456 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.166281 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cad59818d6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.46720687 +0000 UTC m=+0.760494082,LastTimestamp:2026-03-21 00:11:27.46720687 +0000 UTC m=+0.760494082,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.166489 5117 trace.go:236] Trace[1793035767]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Mar-2026 00:11:35.548) (total time: 11618ms): Mar 21 00:11:47 crc kubenswrapper[5117]: Trace[1793035767]: ---"Objects listed" error:services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope 11618ms (00:11:47.166) Mar 21 00:11:47 crc kubenswrapper[5117]: Trace[1793035767]: [11.618252733s] [11.618252733s] END Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.166547 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.166968 5117 trace.go:236] Trace[1166040036]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Mar-2026 00:11:36.115) (total time: 11051ms): Mar 21 00:11:47 crc kubenswrapper[5117]: Trace[1166040036]: ---"Objects listed" error:runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope 11051ms (00:11:47.166) Mar 21 00:11:47 crc kubenswrapper[5117]: Trace[1166040036]: [11.051030065s] [11.051030065s] END Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.167007 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.167719 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadccf8d6e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.58828171 +0000 UTC m=+0.881568882,LastTimestamp:2026-03-21 00:11:27.58828171 +0000 UTC m=+0.881568882,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.168222 5117 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.174126 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd31bcf default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588514767 +0000 UTC m=+0.881801939,LastTimestamp:2026-03-21 00:11:27.588514767 +0000 UTC m=+0.881801939,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.180213 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd36728 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588534056 +0000 UTC m=+0.881821228,LastTimestamp:2026-03-21 00:11:27.588534056 +0000 UTC m=+0.881821228,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.186660 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cae09f53b6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.65223007 +0000 UTC m=+0.945517252,LastTimestamp:2026-03-21 00:11:27.65223007 +0000 UTC m=+0.945517252,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.194678 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadccf8d6e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadccf8d6e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.58828171 +0000 UTC m=+0.881568882,LastTimestamp:2026-03-21 00:11:27.680207072 +0000 UTC m=+0.973494284,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.205851 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd31bcf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd31bcf default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588514767 +0000 UTC m=+0.881801939,LastTimestamp:2026-03-21 00:11:27.680284138 +0000 UTC m=+0.973571350,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.211129 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd36728\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd36728 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588534056 +0000 UTC m=+0.881821228,LastTimestamp:2026-03-21 00:11:27.680307906 +0000 UTC m=+0.973595118,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.216131 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadccf8d6e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadccf8d6e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.58828171 +0000 UTC m=+0.881568882,LastTimestamp:2026-03-21 00:11:27.683033646 +0000 UTC m=+0.976320858,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.221888 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadccf8d6e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadccf8d6e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.58828171 +0000 UTC m=+0.881568882,LastTimestamp:2026-03-21 00:11:27.683074943 +0000 UTC m=+0.976362155,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.228698 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd31bcf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd31bcf default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588514767 +0000 UTC m=+0.881801939,LastTimestamp:2026-03-21 00:11:27.68314127 +0000 UTC m=+0.976428482,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.232611 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd36728\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd36728 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588534056 +0000 UTC m=+0.881821228,LastTimestamp:2026-03-21 00:11:27.683166278 +0000 UTC m=+0.976453490,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.236834 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd31bcf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd31bcf default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588514767 +0000 UTC m=+0.881801939,LastTimestamp:2026-03-21 00:11:27.683195167 +0000 UTC m=+0.976482379,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.241291 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd36728\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd36728 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588534056 +0000 UTC m=+0.881821228,LastTimestamp:2026-03-21 00:11:27.683232525 +0000 UTC m=+0.976519747,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.248103 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadccf8d6e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadccf8d6e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.58828171 +0000 UTC m=+0.881568882,LastTimestamp:2026-03-21 00:11:27.685863239 +0000 UTC m=+0.979150421,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.254579 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd31bcf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd31bcf default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588514767 +0000 UTC m=+0.881801939,LastTimestamp:2026-03-21 00:11:27.685886738 +0000 UTC m=+0.979173920,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.258507 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd36728\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd36728 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588534056 +0000 UTC m=+0.881821228,LastTimestamp:2026-03-21 00:11:27.685901257 +0000 UTC m=+0.979188439,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.262616 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadccf8d6e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadccf8d6e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.58828171 +0000 UTC m=+0.881568882,LastTimestamp:2026-03-21 00:11:27.687597864 +0000 UTC m=+0.980885076,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.269050 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd31bcf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd31bcf default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588514767 +0000 UTC m=+0.881801939,LastTimestamp:2026-03-21 00:11:27.687642391 +0000 UTC m=+0.980929613,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.273058 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd36728\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd36728 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588534056 +0000 UTC m=+0.881821228,LastTimestamp:2026-03-21 00:11:27.68766298 +0000 UTC m=+0.980950192,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.274412 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadccf8d6e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadccf8d6e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.58828171 +0000 UTC m=+0.881568882,LastTimestamp:2026-03-21 00:11:27.687754304 +0000 UTC m=+0.981041486,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.281598 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd31bcf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd31bcf default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588514767 +0000 UTC m=+0.881801939,LastTimestamp:2026-03-21 00:11:27.687773723 +0000 UTC m=+0.981060905,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.288460 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd36728\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd36728 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588534056 +0000 UTC m=+0.881821228,LastTimestamp:2026-03-21 00:11:27.687787472 +0000 UTC m=+0.981074654,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.294381 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadccf8d6e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadccf8d6e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.58828171 +0000 UTC m=+0.881568882,LastTimestamp:2026-03-21 00:11:27.687840189 +0000 UTC m=+0.981127411,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.300547 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189eb2cadcd31bcf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189eb2cadcd31bcf default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:27.588514767 +0000 UTC m=+0.881801939,LastTimestamp:2026-03-21 00:11:27.687935014 +0000 UTC m=+0.981222226,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.307243 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cafa01aec8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.078106312 +0000 UTC m=+1.371393484,LastTimestamp:2026-03-21 00:11:28.078106312 +0000 UTC m=+1.371393484,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.313442 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cafa8537ff openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.086726655 +0000 UTC m=+1.380013827,LastTimestamp:2026-03-21 00:11:28.086726655 +0000 UTC m=+1.380013827,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.317807 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189eb2cafb414605 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.099051013 +0000 UTC m=+1.392338205,LastTimestamp:2026-03-21 00:11:28.099051013 +0000 UTC m=+1.392338205,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.322548 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cafba0dc27 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.105315367 +0000 UTC m=+1.398602549,LastTimestamp:2026-03-21 00:11:28.105315367 +0000 UTC m=+1.398602549,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.326622 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cafba3171d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.105461533 +0000 UTC m=+1.398748745,LastTimestamp:2026-03-21 00:11:28.105461533 +0000 UTC m=+1.398748745,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.330754 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb233970f1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container: wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.769626353 +0000 UTC m=+2.062913525,LastTimestamp:2026-03-21 00:11:28.769626353 +0000 UTC m=+2.062913525,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.335621 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb234731c5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.770527685 +0000 UTC m=+2.063814867,LastTimestamp:2026-03-21 00:11:28.770527685 +0000 UTC m=+2.063814867,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.339802 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb23476627 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container: kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.770541095 +0000 UTC m=+2.063828267,LastTimestamp:2026-03-21 00:11:28.770541095 +0000 UTC m=+2.063828267,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.344739 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cb2350fb83 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.771169155 +0000 UTC m=+2.064456347,LastTimestamp:2026-03-21 00:11:28.771169155 +0000 UTC m=+2.064456347,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.348977 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189eb2cb23583fec openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.77164542 +0000 UTC m=+2.064932592,LastTimestamp:2026-03-21 00:11:28.77164542 +0000 UTC m=+2.064932592,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.353705 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb2434540d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.786068493 +0000 UTC m=+2.079355665,LastTimestamp:2026-03-21 00:11:28.786068493 +0000 UTC m=+2.079355665,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.358565 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cb244355a6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.787051942 +0000 UTC m=+2.080339114,LastTimestamp:2026-03-21 00:11:28.787051942 +0000 UTC m=+2.080339114,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.362632 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb24496bec openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.78745086 +0000 UTC m=+2.080738032,LastTimestamp:2026-03-21 00:11:28.78745086 +0000 UTC m=+2.080738032,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.367293 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb2450d9d0 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.787937744 +0000 UTC m=+2.081224916,LastTimestamp:2026-03-21 00:11:28.787937744 +0000 UTC m=+2.081224916,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.371655 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb24588aa9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.788441769 +0000 UTC m=+2.081728951,LastTimestamp:2026-03-21 00:11:28.788441769 +0000 UTC m=+2.081728951,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.376817 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189eb2cb2458d40e openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:28.788460558 +0000 UTC m=+2.081747740,LastTimestamp:2026-03-21 00:11:28.788460558 +0000 UTC m=+2.081747740,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.381418 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb37f9b9c6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container: cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.11777223 +0000 UTC m=+2.411059392,LastTimestamp:2026-03-21 00:11:29.11777223 +0000 UTC m=+2.411059392,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.386348 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb3884d674 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.126889076 +0000 UTC m=+2.420176258,LastTimestamp:2026-03-21 00:11:29.126889076 +0000 UTC m=+2.420176258,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.392041 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb38a4e7f8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.128990712 +0000 UTC m=+2.422277914,LastTimestamp:2026-03-21 00:11:29.128990712 +0000 UTC m=+2.422277914,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.398416 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb54a8f20e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.599017486 +0000 UTC m=+2.892304678,LastTimestamp:2026-03-21 00:11:29.599017486 +0000 UTC m=+2.892304678,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.403558 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cb54cb3b48 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.601264456 +0000 UTC m=+2.894551628,LastTimestamp:2026-03-21 00:11:29.601264456 +0000 UTC m=+2.894551628,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.412276 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb552fdfe9 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.607860201 +0000 UTC m=+2.901147383,LastTimestamp:2026-03-21 00:11:29.607860201 +0000 UTC m=+2.901147383,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.420812 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189eb2cb55304ece openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.60788859 +0000 UTC m=+2.901175762,LastTimestamp:2026-03-21 00:11:29.60788859 +0000 UTC m=+2.901175762,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.421310 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.426720 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb58ef7129 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container: kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.670746409 +0000 UTC m=+2.964033581,LastTimestamp:2026-03-21 00:11:29.670746409 +0000 UTC m=+2.964033581,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.432079 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb5b2a5aad openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.708161709 +0000 UTC m=+3.001448881,LastTimestamp:2026-03-21 00:11:29.708161709 +0000 UTC m=+3.001448881,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.437616 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb5b43413e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.709793598 +0000 UTC m=+3.003080770,LastTimestamp:2026-03-21 00:11:29.709793598 +0000 UTC m=+3.003080770,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.445865 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cb65f5232a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container: etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.889223466 +0000 UTC m=+3.182510638,LastTimestamp:2026-03-21 00:11:29.889223466 +0000 UTC m=+3.182510638,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.450088 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb65f6ed53 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container: kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.889340755 +0000 UTC m=+3.182627927,LastTimestamp:2026-03-21 00:11:29.889340755 +0000 UTC m=+3.182627927,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.455738 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189eb2cb65f7f530 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container: kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.889408304 +0000 UTC m=+3.182695476,LastTimestamp:2026-03-21 00:11:29.889408304 +0000 UTC m=+3.182695476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.459530 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb65f888eb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container: kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.889446123 +0000 UTC m=+3.182733295,LastTimestamp:2026-03-21 00:11:29.889446123 +0000 UTC m=+3.182733295,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.463909 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb671649cd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.908173261 +0000 UTC m=+3.201460433,LastTimestamp:2026-03-21 00:11:29.908173261 +0000 UTC m=+3.201460433,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.468669 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb6731c99c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.909975452 +0000 UTC m=+3.203262624,LastTimestamp:2026-03-21 00:11:29.909975452 +0000 UTC m=+3.203262624,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.472923 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb6787937b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.915597691 +0000 UTC m=+3.208884863,LastTimestamp:2026-03-21 00:11:29.915597691 +0000 UTC m=+3.208884863,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.478484 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189eb2cb6787bdf1 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.915608561 +0000 UTC m=+3.208895733,LastTimestamp:2026-03-21 00:11:29.915608561 +0000 UTC m=+3.208895733,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.483241 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb67955452 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.916499026 +0000 UTC m=+3.209786198,LastTimestamp:2026-03-21 00:11:29.916499026 +0000 UTC m=+3.209786198,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.500499 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cb67b3a11d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:29.918484765 +0000 UTC m=+3.211771937,LastTimestamp:2026-03-21 00:11:29.918484765 +0000 UTC m=+3.211771937,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.507182 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb6e32f293 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container: kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.027491987 +0000 UTC m=+3.320779159,LastTimestamp:2026-03-21 00:11:30.027491987 +0000 UTC m=+3.320779159,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.511980 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cb6f914f75 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.050453365 +0000 UTC m=+3.343740537,LastTimestamp:2026-03-21 00:11:30.050453365 +0000 UTC m=+3.343740537,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.520158 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb74ca2a69 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container: kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.138065513 +0000 UTC m=+3.431352685,LastTimestamp:2026-03-21 00:11:30.138065513 +0000 UTC m=+3.431352685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.531978 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb750de85c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container: kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.142505052 +0000 UTC m=+3.435792224,LastTimestamp:2026-03-21 00:11:30.142505052 +0000 UTC m=+3.435792224,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.536056 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb75d65d57 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.155642199 +0000 UTC m=+3.448929371,LastTimestamp:2026-03-21 00:11:30.155642199 +0000 UTC m=+3.448929371,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.541472 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb75e9ff50 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.156928848 +0000 UTC m=+3.450216020,LastTimestamp:2026-03-21 00:11:30.156928848 +0000 UTC m=+3.450216020,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.546365 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb7622a577 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.160641399 +0000 UTC m=+3.453928571,LastTimestamp:2026-03-21 00:11:30.160641399 +0000 UTC m=+3.453928571,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.550151 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb7632aa60 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.161691232 +0000 UTC m=+3.454978394,LastTimestamp:2026-03-21 00:11:30.161691232 +0000 UTC m=+3.454978394,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.553296 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb815e3bc7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container: kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.349095879 +0000 UTC m=+3.642383051,LastTimestamp:2026-03-21 00:11:30.349095879 +0000 UTC m=+3.642383051,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.557862 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb81772aae openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container: kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.350729902 +0000 UTC m=+3.644017074,LastTimestamp:2026-03-21 00:11:30.350729902 +0000 UTC m=+3.644017074,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.562032 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb823d86eb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.363729643 +0000 UTC m=+3.657016825,LastTimestamp:2026-03-21 00:11:30.363729643 +0000 UTC m=+3.657016825,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.565700 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb825aca70 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.365647472 +0000 UTC m=+3.658934644,LastTimestamp:2026-03-21 00:11:30.365647472 +0000 UTC m=+3.658934644,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.569938 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189eb2cb82905d10 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.369158416 +0000 UTC m=+3.662445608,LastTimestamp:2026-03-21 00:11:30.369158416 +0000 UTC m=+3.662445608,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.573769 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb8f73fa15 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container: kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.585401877 +0000 UTC m=+3.878689049,LastTimestamp:2026-03-21 00:11:30.585401877 +0000 UTC m=+3.878689049,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.577397 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb909be936 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.604796214 +0000 UTC m=+3.898083386,LastTimestamp:2026-03-21 00:11:30.604796214 +0000 UTC m=+3.898083386,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.578999 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb90ad57ea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.605938666 +0000 UTC m=+3.899225838,LastTimestamp:2026-03-21 00:11:30.605938666 +0000 UTC m=+3.899225838,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.581135 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cb9137080f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.614962191 +0000 UTC m=+3.908249363,LastTimestamp:2026-03-21 00:11:30.614962191 +0000 UTC m=+3.908249363,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.584060 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb9cbb3302 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container: kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.808173314 +0000 UTC m=+4.101460486,LastTimestamp:2026-03-21 00:11:30.808173314 +0000 UTC m=+4.101460486,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.585895 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb9e291725 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.832152357 +0000 UTC m=+4.125439529,LastTimestamp:2026-03-21 00:11:30.832152357 +0000 UTC m=+4.125439529,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.590065 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cb9e999443 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container: etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.839524419 +0000 UTC m=+4.132811591,LastTimestamp:2026-03-21 00:11:30.839524419 +0000 UTC m=+4.132811591,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.593910 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cb9f9a9845 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.856368197 +0000 UTC m=+4.149655369,LastTimestamp:2026-03-21 00:11:30.856368197 +0000 UTC m=+4.149655369,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.600037 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbcecab420 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:31.648050208 +0000 UTC m=+4.941337380,LastTimestamp:2026-03-21 00:11:31.648050208 +0000 UTC m=+4.941337380,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.604422 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbdd3e2743 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container: etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:31.890497347 +0000 UTC m=+5.183784519,LastTimestamp:2026-03-21 00:11:31.890497347 +0000 UTC m=+5.183784519,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.609502 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbddd4e260 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:31.900375648 +0000 UTC m=+5.193662820,LastTimestamp:2026-03-21 00:11:31.900375648 +0000 UTC m=+5.193662820,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.614196 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbdde13f7a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:31.901185914 +0000 UTC m=+5.194473076,LastTimestamp:2026-03-21 00:11:31.901185914 +0000 UTC m=+5.194473076,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.617928 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbead48c88 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container: etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.11845748 +0000 UTC m=+5.411744652,LastTimestamp:2026-03-21 00:11:32.11845748 +0000 UTC m=+5.411744652,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.622518 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbebc26b36 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.134046518 +0000 UTC m=+5.427333710,LastTimestamp:2026-03-21 00:11:32.134046518 +0000 UTC m=+5.427333710,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.626454 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbebd66f00 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.135358208 +0000 UTC m=+5.428645390,LastTimestamp:2026-03-21 00:11:32.135358208 +0000 UTC m=+5.428645390,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.629986 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbf6a9c71a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container: etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.316981018 +0000 UTC m=+5.610268190,LastTimestamp:2026-03-21 00:11:32.316981018 +0000 UTC m=+5.610268190,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.634203 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbf763e873 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.329179251 +0000 UTC m=+5.622466423,LastTimestamp:2026-03-21 00:11:32.329179251 +0000 UTC m=+5.622466423,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.638655 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cbf772f169 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.330164585 +0000 UTC m=+5.623451757,LastTimestamp:2026-03-21 00:11:32.330164585 +0000 UTC m=+5.623451757,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.643108 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cc02d346eb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container: etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.521027307 +0000 UTC m=+5.814314469,LastTimestamp:2026-03-21 00:11:32.521027307 +0000 UTC m=+5.814314469,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.646753 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cc035bdb96 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.529978262 +0000 UTC m=+5.823265434,LastTimestamp:2026-03-21 00:11:32.529978262 +0000 UTC m=+5.823265434,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.650678 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cc0373ee6b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.531555947 +0000 UTC m=+5.824843149,LastTimestamp:2026-03-21 00:11:32.531555947 +0000 UTC m=+5.824843149,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.654361 5117 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.655041 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cc1014d19c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container: etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.74342646 +0000 UTC m=+6.036713622,LastTimestamp:2026-03-21 00:11:32.74342646 +0000 UTC m=+6.036713622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.660198 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189eb2cc10fc5dcd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:32.758601165 +0000 UTC m=+6.051888377,LastTimestamp:2026-03-21 00:11:32.758601165 +0000 UTC m=+6.051888377,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.665655 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 21 00:11:47 crc kubenswrapper[5117]: &Event{ObjectMeta:{kube-controller-manager-crc.189eb2cd8005f6cf openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://localhost:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 21 00:11:47 crc kubenswrapper[5117]: body: Mar 21 00:11:47 crc kubenswrapper[5117]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:38.916468431 +0000 UTC m=+12.209755643,LastTimestamp:2026-03-21 00:11:38.916468431 +0000 UTC m=+12.209755643,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 00:11:47 crc kubenswrapper[5117]: > Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.669153 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189eb2cd80088162 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:38.916634978 +0000 UTC m=+12.209922190,LastTimestamp:2026-03-21 00:11:38.916634978 +0000 UTC m=+12.209922190,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.673208 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 21 00:11:47 crc kubenswrapper[5117]: &Event{ObjectMeta:{kube-apiserver-crc.189eb2ce384e7443 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:6443/livez": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 21 00:11:47 crc kubenswrapper[5117]: body: Mar 21 00:11:47 crc kubenswrapper[5117]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:42.008226883 +0000 UTC m=+15.301514095,LastTimestamp:2026-03-21 00:11:42.008226883 +0000 UTC m=+15.301514095,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 00:11:47 crc kubenswrapper[5117]: > Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.677113 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.677155 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ce3850640f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:42.008353807 +0000 UTC m=+15.301641019,LastTimestamp:2026-03-21 00:11:42.008353807 +0000 UTC m=+15.301641019,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.677351 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.678118 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.678153 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.678166 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.678535 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.681808 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 21 00:11:47 crc kubenswrapper[5117]: &Event{ObjectMeta:{kube-apiserver-crc.189eb2ce41bdf543 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 21 00:11:47 crc kubenswrapper[5117]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 21 00:11:47 crc kubenswrapper[5117]: Mar 21 00:11:47 crc kubenswrapper[5117]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:42.166529347 +0000 UTC m=+15.459816539,LastTimestamp:2026-03-21 00:11:42.166529347 +0000 UTC m=+15.459816539,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 21 00:11:47 crc kubenswrapper[5117]: > Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.686331 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ce41bf2de3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:42.166609379 +0000 UTC m=+15.459896571,LastTimestamp:2026-03-21 00:11:42.166609379 +0000 UTC m=+15.459896571,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.690422 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2cb90ad57ea\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb90ad57ea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.605938666 +0000 UTC m=+3.899225838,LastTimestamp:2026-03-21 00:11:42.694235134 +0000 UTC m=+15.987522336,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.693890 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2cb9cbb3302\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb9cbb3302 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container: kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.808173314 +0000 UTC m=+4.101460486,LastTimestamp:2026-03-21 00:11:42.927467102 +0000 UTC m=+16.220754284,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.697209 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2cb9e291725\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb9e291725 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.832152357 +0000 UTC m=+4.125439529,LastTimestamp:2026-03-21 00:11:42.963230841 +0000 UTC m=+16.256518023,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.701070 5117 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ced93264ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,LastTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.704485 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2ced93264ae\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ced93264ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,LastTimestamp:2026-03-21 00:11:45.713744303 +0000 UTC m=+19.007031485,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.708459 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2ced93264ae\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ced93264ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,LastTimestamp:2026-03-21 00:11:47.020224432 +0000 UTC m=+20.313511654,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.715814 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.716341 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.716376 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.716387 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.716708 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:47 crc kubenswrapper[5117]: I0321 00:11:47.716944 5117 scope.go:117] "RemoveContainer" containerID="fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.717122 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:11:47 crc kubenswrapper[5117]: E0321 00:11:47.720469 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2ced93264ae\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ced93264ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,LastTimestamp:2026-03-21 00:11:47.717095717 +0000 UTC m=+21.010382889,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.192403 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.192597 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.193874 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.193942 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.193967 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:48 crc kubenswrapper[5117]: E0321 00:11:48.194499 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.197937 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.417980 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.708520 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.708724 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.709431 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.709461 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.709470 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:48 crc kubenswrapper[5117]: E0321 00:11:48.709878 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.717865 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.718230 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.718325 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.718340 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:48 crc kubenswrapper[5117]: E0321 00:11:48.718569 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.722689 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.722842 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.723461 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.723492 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:48 crc kubenswrapper[5117]: I0321 00:11:48.723504 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:48 crc kubenswrapper[5117]: E0321 00:11:48.723792 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:49 crc kubenswrapper[5117]: I0321 00:11:49.419450 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:50 crc kubenswrapper[5117]: E0321 00:11:50.091690 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 00:11:50 crc kubenswrapper[5117]: I0321 00:11:50.418565 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:51 crc kubenswrapper[5117]: I0321 00:11:51.417988 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:52 crc kubenswrapper[5117]: I0321 00:11:52.421868 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.422593 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.569673 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.571489 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.571583 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.571612 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.571667 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:11:53 crc kubenswrapper[5117]: E0321 00:11:53.587054 5117 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.701304 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.701775 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.703200 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.703276 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.703288 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:11:53 crc kubenswrapper[5117]: E0321 00:11:53.703746 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:11:53 crc kubenswrapper[5117]: I0321 00:11:53.704049 5117 scope.go:117] "RemoveContainer" containerID="fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2" Mar 21 00:11:53 crc kubenswrapper[5117]: E0321 00:11:53.704359 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:11:53 crc kubenswrapper[5117]: E0321 00:11:53.712604 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2ced93264ae\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ced93264ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,LastTimestamp:2026-03-21 00:11:53.704327134 +0000 UTC m=+26.997614306,Count:5,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:11:54 crc kubenswrapper[5117]: I0321 00:11:54.421634 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:55 crc kubenswrapper[5117]: I0321 00:11:55.422036 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:56 crc kubenswrapper[5117]: I0321 00:11:56.420450 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:56 crc kubenswrapper[5117]: E0321 00:11:56.477395 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 21 00:11:57 crc kubenswrapper[5117]: E0321 00:11:57.096968 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 00:11:57 crc kubenswrapper[5117]: E0321 00:11:57.331098 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 21 00:11:57 crc kubenswrapper[5117]: I0321 00:11:57.420189 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:57 crc kubenswrapper[5117]: E0321 00:11:57.654709 5117 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 00:11:58 crc kubenswrapper[5117]: I0321 00:11:58.422395 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:11:59 crc kubenswrapper[5117]: E0321 00:11:59.249339 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 21 00:11:59 crc kubenswrapper[5117]: E0321 00:11:59.374556 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 21 00:11:59 crc kubenswrapper[5117]: I0321 00:11:59.419780 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:00 crc kubenswrapper[5117]: I0321 00:12:00.421939 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:00 crc kubenswrapper[5117]: I0321 00:12:00.588041 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:00 crc kubenswrapper[5117]: I0321 00:12:00.589880 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:00 crc kubenswrapper[5117]: I0321 00:12:00.590166 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:00 crc kubenswrapper[5117]: I0321 00:12:00.590420 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:00 crc kubenswrapper[5117]: I0321 00:12:00.590615 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:12:00 crc kubenswrapper[5117]: E0321 00:12:00.600760 5117 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 00:12:01 crc kubenswrapper[5117]: I0321 00:12:01.423923 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:03 crc kubenswrapper[5117]: I0321 00:12:03.158468 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:03 crc kubenswrapper[5117]: I0321 00:12:03.420008 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:04 crc kubenswrapper[5117]: E0321 00:12:04.105707 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 00:12:04 crc kubenswrapper[5117]: I0321 00:12:04.420956 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:05 crc kubenswrapper[5117]: I0321 00:12:05.419563 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:05 crc kubenswrapper[5117]: I0321 00:12:05.578388 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:05 crc kubenswrapper[5117]: I0321 00:12:05.579857 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:05 crc kubenswrapper[5117]: I0321 00:12:05.579950 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:05 crc kubenswrapper[5117]: I0321 00:12:05.579976 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:05 crc kubenswrapper[5117]: E0321 00:12:05.580678 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:05 crc kubenswrapper[5117]: I0321 00:12:05.581202 5117 scope.go:117] "RemoveContainer" containerID="fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2" Mar 21 00:12:05 crc kubenswrapper[5117]: E0321 00:12:05.590548 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2cb90ad57ea\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb90ad57ea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.605938666 +0000 UTC m=+3.899225838,LastTimestamp:2026-03-21 00:12:05.58246576 +0000 UTC m=+38.875752932,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:12:05 crc kubenswrapper[5117]: E0321 00:12:05.802184 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2cb9cbb3302\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb9cbb3302 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container: kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.808173314 +0000 UTC m=+4.101460486,LastTimestamp:2026-03-21 00:12:05.796993883 +0000 UTC m=+39.090281055,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:12:05 crc kubenswrapper[5117]: E0321 00:12:05.816111 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2cb9e291725\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2cb9e291725 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:30.832152357 +0000 UTC m=+4.125439529,LastTimestamp:2026-03-21 00:12:05.809792096 +0000 UTC m=+39.103079268,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.419866 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.775967 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.776791 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.778777 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="f9fd30d306977056d6f46ef73f30465af70e06f0ab642bcb553568eb1c551d5b" exitCode=255 Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.778860 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"f9fd30d306977056d6f46ef73f30465af70e06f0ab642bcb553568eb1c551d5b"} Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.778922 5117 scope.go:117] "RemoveContainer" containerID="fde027824fca9a8ce24a3faf7150c11975f6fab20ce1450a66b51a545d9449b2" Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.779258 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.780198 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.780263 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.780311 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:06 crc kubenswrapper[5117]: E0321 00:12:06.781127 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:06 crc kubenswrapper[5117]: I0321 00:12:06.781589 5117 scope.go:117] "RemoveContainer" containerID="f9fd30d306977056d6f46ef73f30465af70e06f0ab642bcb553568eb1c551d5b" Mar 21 00:12:06 crc kubenswrapper[5117]: E0321 00:12:06.782018 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:12:06 crc kubenswrapper[5117]: E0321 00:12:06.789318 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2ced93264ae\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ced93264ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,LastTimestamp:2026-03-21 00:12:06.7819278 +0000 UTC m=+40.075215002,Count:6,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:12:07 crc kubenswrapper[5117]: I0321 00:12:07.416881 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:07 crc kubenswrapper[5117]: I0321 00:12:07.601345 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:07 crc kubenswrapper[5117]: I0321 00:12:07.602576 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:07 crc kubenswrapper[5117]: I0321 00:12:07.602617 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:07 crc kubenswrapper[5117]: I0321 00:12:07.602632 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:07 crc kubenswrapper[5117]: I0321 00:12:07.602659 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:12:07 crc kubenswrapper[5117]: E0321 00:12:07.618528 5117 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 00:12:07 crc kubenswrapper[5117]: E0321 00:12:07.655747 5117 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 00:12:07 crc kubenswrapper[5117]: I0321 00:12:07.783382 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Mar 21 00:12:08 crc kubenswrapper[5117]: I0321 00:12:08.421778 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:09 crc kubenswrapper[5117]: I0321 00:12:09.422089 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:10 crc kubenswrapper[5117]: I0321 00:12:10.418470 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:11 crc kubenswrapper[5117]: E0321 00:12:11.114327 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 00:12:11 crc kubenswrapper[5117]: I0321 00:12:11.420114 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:12 crc kubenswrapper[5117]: I0321 00:12:12.427297 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:13 crc kubenswrapper[5117]: I0321 00:12:13.421411 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:13 crc kubenswrapper[5117]: I0321 00:12:13.701098 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:12:13 crc kubenswrapper[5117]: I0321 00:12:13.701691 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:13 crc kubenswrapper[5117]: I0321 00:12:13.703098 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:13 crc kubenswrapper[5117]: I0321 00:12:13.703156 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:13 crc kubenswrapper[5117]: I0321 00:12:13.703174 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:13 crc kubenswrapper[5117]: E0321 00:12:13.703716 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:13 crc kubenswrapper[5117]: I0321 00:12:13.704163 5117 scope.go:117] "RemoveContainer" containerID="f9fd30d306977056d6f46ef73f30465af70e06f0ab642bcb553568eb1c551d5b" Mar 21 00:12:13 crc kubenswrapper[5117]: E0321 00:12:13.704525 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:12:13 crc kubenswrapper[5117]: E0321 00:12:13.712042 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2ced93264ae\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ced93264ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,LastTimestamp:2026-03-21 00:12:13.7044778 +0000 UTC m=+46.997765002,Count:7,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:12:14 crc kubenswrapper[5117]: I0321 00:12:14.422756 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:14 crc kubenswrapper[5117]: I0321 00:12:14.618978 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:14 crc kubenswrapper[5117]: I0321 00:12:14.620477 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:14 crc kubenswrapper[5117]: I0321 00:12:14.620565 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:14 crc kubenswrapper[5117]: I0321 00:12:14.620588 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:14 crc kubenswrapper[5117]: I0321 00:12:14.620635 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:12:14 crc kubenswrapper[5117]: E0321 00:12:14.635196 5117 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 00:12:15 crc kubenswrapper[5117]: I0321 00:12:15.099377 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:12:15 crc kubenswrapper[5117]: I0321 00:12:15.100173 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:15 crc kubenswrapper[5117]: I0321 00:12:15.102991 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:15 crc kubenswrapper[5117]: I0321 00:12:15.103082 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:15 crc kubenswrapper[5117]: I0321 00:12:15.103111 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:15 crc kubenswrapper[5117]: E0321 00:12:15.103908 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:15 crc kubenswrapper[5117]: I0321 00:12:15.104556 5117 scope.go:117] "RemoveContainer" containerID="f9fd30d306977056d6f46ef73f30465af70e06f0ab642bcb553568eb1c551d5b" Mar 21 00:12:15 crc kubenswrapper[5117]: E0321 00:12:15.105011 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:12:15 crc kubenswrapper[5117]: E0321 00:12:15.114619 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2ced93264ae\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ced93264ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,LastTimestamp:2026-03-21 00:12:15.104944217 +0000 UTC m=+48.398231429,Count:8,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:12:15 crc kubenswrapper[5117]: I0321 00:12:15.422061 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:15 crc kubenswrapper[5117]: E0321 00:12:15.844956 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 21 00:12:16 crc kubenswrapper[5117]: I0321 00:12:16.422468 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:17 crc kubenswrapper[5117]: E0321 00:12:17.006989 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 21 00:12:17 crc kubenswrapper[5117]: I0321 00:12:17.210961 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:12:17 crc kubenswrapper[5117]: I0321 00:12:17.211202 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:17 crc kubenswrapper[5117]: I0321 00:12:17.212241 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:17 crc kubenswrapper[5117]: I0321 00:12:17.212305 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:17 crc kubenswrapper[5117]: I0321 00:12:17.212317 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:17 crc kubenswrapper[5117]: E0321 00:12:17.212682 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:17 crc kubenswrapper[5117]: I0321 00:12:17.422213 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:17 crc kubenswrapper[5117]: E0321 00:12:17.537684 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 21 00:12:17 crc kubenswrapper[5117]: E0321 00:12:17.657034 5117 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 00:12:18 crc kubenswrapper[5117]: E0321 00:12:18.123770 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 00:12:18 crc kubenswrapper[5117]: I0321 00:12:18.421796 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:19 crc kubenswrapper[5117]: I0321 00:12:19.420372 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:20 crc kubenswrapper[5117]: I0321 00:12:20.419649 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:20 crc kubenswrapper[5117]: E0321 00:12:20.767550 5117 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 21 00:12:21 crc kubenswrapper[5117]: I0321 00:12:21.422461 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:21 crc kubenswrapper[5117]: I0321 00:12:21.636174 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:21 crc kubenswrapper[5117]: I0321 00:12:21.638177 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:21 crc kubenswrapper[5117]: I0321 00:12:21.638235 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:21 crc kubenswrapper[5117]: I0321 00:12:21.638285 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:21 crc kubenswrapper[5117]: I0321 00:12:21.638326 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:12:21 crc kubenswrapper[5117]: E0321 00:12:21.654541 5117 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 00:12:22 crc kubenswrapper[5117]: I0321 00:12:22.420518 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:23 crc kubenswrapper[5117]: I0321 00:12:23.420560 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:24 crc kubenswrapper[5117]: I0321 00:12:24.421696 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:25 crc kubenswrapper[5117]: E0321 00:12:25.133719 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 00:12:25 crc kubenswrapper[5117]: I0321 00:12:25.420576 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:26 crc kubenswrapper[5117]: I0321 00:12:26.422477 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:26 crc kubenswrapper[5117]: I0321 00:12:26.579408 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:26 crc kubenswrapper[5117]: I0321 00:12:26.580940 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:26 crc kubenswrapper[5117]: I0321 00:12:26.581244 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:26 crc kubenswrapper[5117]: I0321 00:12:26.581293 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:26 crc kubenswrapper[5117]: E0321 00:12:26.581919 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:26 crc kubenswrapper[5117]: I0321 00:12:26.582492 5117 scope.go:117] "RemoveContainer" containerID="f9fd30d306977056d6f46ef73f30465af70e06f0ab642bcb553568eb1c551d5b" Mar 21 00:12:26 crc kubenswrapper[5117]: E0321 00:12:26.582913 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:12:26 crc kubenswrapper[5117]: E0321 00:12:26.591792 5117 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189eb2ced93264ae\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189eb2ced93264ae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:11:44.707519662 +0000 UTC m=+18.000806834,LastTimestamp:2026-03-21 00:12:26.582853452 +0000 UTC m=+59.876140654,Count:9,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:12:27 crc kubenswrapper[5117]: I0321 00:12:27.420293 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:27 crc kubenswrapper[5117]: E0321 00:12:27.657479 5117 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 00:12:28 crc kubenswrapper[5117]: I0321 00:12:28.421790 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:28 crc kubenswrapper[5117]: I0321 00:12:28.655472 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:28 crc kubenswrapper[5117]: I0321 00:12:28.657071 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:28 crc kubenswrapper[5117]: I0321 00:12:28.657164 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:28 crc kubenswrapper[5117]: I0321 00:12:28.657188 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:28 crc kubenswrapper[5117]: I0321 00:12:28.657646 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:12:28 crc kubenswrapper[5117]: E0321 00:12:28.674188 5117 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 21 00:12:29 crc kubenswrapper[5117]: I0321 00:12:29.421994 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:30 crc kubenswrapper[5117]: I0321 00:12:30.422089 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:31 crc kubenswrapper[5117]: I0321 00:12:31.421686 5117 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 21 00:12:32 crc kubenswrapper[5117]: E0321 00:12:32.140730 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 21 00:12:32 crc kubenswrapper[5117]: I0321 00:12:32.153929 5117 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-5gq7w" Mar 21 00:12:32 crc kubenswrapper[5117]: I0321 00:12:32.161101 5117 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-5gq7w" Mar 21 00:12:32 crc kubenswrapper[5117]: I0321 00:12:32.185274 5117 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 21 00:12:32 crc kubenswrapper[5117]: I0321 00:12:32.217667 5117 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 21 00:12:33 crc kubenswrapper[5117]: I0321 00:12:33.163423 5117 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2026-04-20 00:07:32 +0000 UTC" deadline="2026-04-14 21:16:08.832940964 +0000 UTC" Mar 21 00:12:33 crc kubenswrapper[5117]: I0321 00:12:33.163474 5117 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="597h3m35.66947044s" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.674788 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.676527 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.676594 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.676617 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.676818 5117 kubelet_node_status.go:78] "Attempting to register node" node="crc" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.686649 5117 kubelet_node_status.go:127] "Node was previously registered" node="crc" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.687099 5117 kubelet_node_status.go:81] "Successfully registered node" node="crc" Mar 21 00:12:35 crc kubenswrapper[5117]: E0321 00:12:35.687145 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.691202 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.691312 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.691346 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.691380 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.691407 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:35Z","lastTransitionTime":"2026-03-21T00:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:35 crc kubenswrapper[5117]: E0321 00:12:35.709973 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6152d979-425d-4df2-909b-43e70deffa85\\\",\\\"systemUUID\\\":\\\"a07d7eeb-81b3-459f-bb75-0975d4964835\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.719337 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.719394 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.719409 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.719428 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.719443 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:35Z","lastTransitionTime":"2026-03-21T00:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:35 crc kubenswrapper[5117]: E0321 00:12:35.735537 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6152d979-425d-4df2-909b-43e70deffa85\\\",\\\"systemUUID\\\":\\\"a07d7eeb-81b3-459f-bb75-0975d4964835\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.743976 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.744041 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.744061 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.744098 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.744121 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:35Z","lastTransitionTime":"2026-03-21T00:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:35 crc kubenswrapper[5117]: E0321 00:12:35.757031 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6152d979-425d-4df2-909b-43e70deffa85\\\",\\\"systemUUID\\\":\\\"a07d7eeb-81b3-459f-bb75-0975d4964835\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.773024 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.773090 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.773105 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.773123 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:35 crc kubenswrapper[5117]: I0321 00:12:35.773135 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:35Z","lastTransitionTime":"2026-03-21T00:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:35 crc kubenswrapper[5117]: E0321 00:12:35.785721 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6152d979-425d-4df2-909b-43e70deffa85\\\",\\\"systemUUID\\\":\\\"a07d7eeb-81b3-459f-bb75-0975d4964835\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:35 crc kubenswrapper[5117]: E0321 00:12:35.785967 5117 kubelet_node_status.go:584] "Unable to update node status" err="update node status exceeds retry count" Mar 21 00:12:35 crc kubenswrapper[5117]: E0321 00:12:35.786012 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:35 crc kubenswrapper[5117]: E0321 00:12:35.886435 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:35 crc kubenswrapper[5117]: E0321 00:12:35.987345 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.088147 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.188978 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.289609 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.389815 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.490448 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.591388 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.692344 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.793454 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.894361 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:36 crc kubenswrapper[5117]: E0321 00:12:36.995376 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.096551 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.197360 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.298031 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.398409 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.499101 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.600036 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.658161 5117 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.700448 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.800863 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:37 crc kubenswrapper[5117]: E0321 00:12:37.902081 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.002872 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.103272 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.203859 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.304395 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.405405 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.505722 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.606212 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.706658 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.807345 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:38 crc kubenswrapper[5117]: E0321 00:12:38.908318 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.008824 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.109226 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.209439 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.309886 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.410738 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.511108 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.612316 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.713320 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.813457 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:39 crc kubenswrapper[5117]: E0321 00:12:39.914164 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.014739 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.115297 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.216225 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.317243 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.418181 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.519044 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.619921 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.720301 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.820988 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:40 crc kubenswrapper[5117]: E0321 00:12:40.921809 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.022619 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.123165 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.224207 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.324685 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.424806 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.524992 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: I0321 00:12:41.624592 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.625859 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: I0321 00:12:41.626589 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:41 crc kubenswrapper[5117]: I0321 00:12:41.626892 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:41 crc kubenswrapper[5117]: I0321 00:12:41.627166 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.628154 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:41 crc kubenswrapper[5117]: I0321 00:12:41.628814 5117 scope.go:117] "RemoveContainer" containerID="f9fd30d306977056d6f46ef73f30465af70e06f0ab642bcb553568eb1c551d5b" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.726583 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.827149 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:41 crc kubenswrapper[5117]: I0321 00:12:41.886224 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Mar 21 00:12:41 crc kubenswrapper[5117]: I0321 00:12:41.888695 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17"} Mar 21 00:12:41 crc kubenswrapper[5117]: E0321 00:12:41.927328 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.027435 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.128016 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.229123 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.329788 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.430603 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.531034 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.631571 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.732795 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.832968 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:42 crc kubenswrapper[5117]: I0321 00:12:42.892586 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:42 crc kubenswrapper[5117]: I0321 00:12:42.893605 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:42 crc kubenswrapper[5117]: I0321 00:12:42.893663 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:42 crc kubenswrapper[5117]: I0321 00:12:42.893682 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.894339 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:42 crc kubenswrapper[5117]: E0321 00:12:42.933526 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.034578 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.135085 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.235460 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.336052 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.437020 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.537714 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.638593 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.738792 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.839539 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.898342 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.899410 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.902958 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17" exitCode=255 Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.903075 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17"} Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.903163 5117 scope.go:117] "RemoveContainer" containerID="f9fd30d306977056d6f46ef73f30465af70e06f0ab642bcb553568eb1c551d5b" Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.903522 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.904750 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.904838 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.904854 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.909136 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:43 crc kubenswrapper[5117]: I0321 00:12:43.909723 5117 scope.go:117] "RemoveContainer" containerID="79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.910114 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:12:43 crc kubenswrapper[5117]: E0321 00:12:43.940479 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.041283 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.141773 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.242345 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.342898 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.443489 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.544342 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.644791 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.745231 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.846383 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:44 crc kubenswrapper[5117]: I0321 00:12:44.907635 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Mar 21 00:12:44 crc kubenswrapper[5117]: E0321 00:12:44.946728 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.047234 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: I0321 00:12:45.095458 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:12:45 crc kubenswrapper[5117]: I0321 00:12:45.095840 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:45 crc kubenswrapper[5117]: I0321 00:12:45.097107 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:45 crc kubenswrapper[5117]: I0321 00:12:45.097155 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:45 crc kubenswrapper[5117]: I0321 00:12:45.097177 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.097996 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:45 crc kubenswrapper[5117]: I0321 00:12:45.098414 5117 scope.go:117] "RemoveContainer" containerID="79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.098797 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.148450 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.249068 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.349766 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.450492 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.550868 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.651766 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.752870 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.853280 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:45 crc kubenswrapper[5117]: E0321 00:12:45.954428 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.025506 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.031935 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.031995 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.032014 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.032041 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.032059 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:46Z","lastTransitionTime":"2026-03-21T00:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.049966 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6152d979-425d-4df2-909b-43e70deffa85\\\",\\\"systemUUID\\\":\\\"a07d7eeb-81b3-459f-bb75-0975d4964835\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.061733 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.061797 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.061818 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.061842 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.061861 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:46Z","lastTransitionTime":"2026-03-21T00:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.078405 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6152d979-425d-4df2-909b-43e70deffa85\\\",\\\"systemUUID\\\":\\\"a07d7eeb-81b3-459f-bb75-0975d4964835\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.090136 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.090217 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.090244 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.090299 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.090317 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:46Z","lastTransitionTime":"2026-03-21T00:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.105020 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6152d979-425d-4df2-909b-43e70deffa85\\\",\\\"systemUUID\\\":\\\"a07d7eeb-81b3-459f-bb75-0975d4964835\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.116695 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.116785 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.116812 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.116851 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:46 crc kubenswrapper[5117]: I0321 00:12:46.116877 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:46Z","lastTransitionTime":"2026-03-21T00:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.134143 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6152d979-425d-4df2-909b-43e70deffa85\\\",\\\"systemUUID\\\":\\\"a07d7eeb-81b3-459f-bb75-0975d4964835\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.134465 5117 kubelet_node_status.go:584] "Unable to update node status" err="update node status exceeds retry count" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.134512 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.235030 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.336133 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.436364 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.537366 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.638085 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.738330 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.839555 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:46 crc kubenswrapper[5117]: E0321 00:12:46.940608 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.041566 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.142147 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.242409 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.342818 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.443915 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.544747 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.645881 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.659372 5117 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.746599 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.847079 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:47 crc kubenswrapper[5117]: E0321 00:12:47.947643 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.048238 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.149157 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.250010 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.350693 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.451275 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.552162 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.652904 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.753223 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.854333 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:48 crc kubenswrapper[5117]: E0321 00:12:48.954965 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.056121 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.156242 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.256772 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.357827 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.457906 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.558758 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.659890 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.760434 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.860711 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:49 crc kubenswrapper[5117]: I0321 00:12:49.951670 5117 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Mar 21 00:12:49 crc kubenswrapper[5117]: E0321 00:12:49.961162 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.061858 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.162227 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.262902 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.363216 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.463592 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.564136 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.664418 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.765654 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.866367 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:50 crc kubenswrapper[5117]: E0321 00:12:50.967351 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.067515 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.167964 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.268491 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.369218 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.470203 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.570462 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.671401 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.772309 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.873297 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:51 crc kubenswrapper[5117]: E0321 00:12:51.973533 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.074557 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.175146 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.275405 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: I0321 00:12:52.317699 5117 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Mar 21 00:12:52 crc kubenswrapper[5117]: I0321 00:12:52.336645 5117 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.375780 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.476231 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.577301 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.678314 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.778385 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.878495 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:52 crc kubenswrapper[5117]: I0321 00:12:52.892846 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:12:52 crc kubenswrapper[5117]: I0321 00:12:52.893245 5117 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Mar 21 00:12:52 crc kubenswrapper[5117]: I0321 00:12:52.894511 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:52 crc kubenswrapper[5117]: I0321 00:12:52.894582 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:52 crc kubenswrapper[5117]: I0321 00:12:52.894595 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.895238 5117 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Mar 21 00:12:52 crc kubenswrapper[5117]: I0321 00:12:52.895606 5117 scope.go:117] "RemoveContainer" containerID="79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.895857 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:12:52 crc kubenswrapper[5117]: E0321 00:12:52.979459 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:53 crc kubenswrapper[5117]: E0321 00:12:53.080268 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:53 crc kubenswrapper[5117]: E0321 00:12:53.180941 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:53 crc kubenswrapper[5117]: E0321 00:12:53.281891 5117 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.347870 5117 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.376723 5117 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.383960 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.384016 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.384038 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.384068 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.384087 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:53Z","lastTransitionTime":"2026-03-21T00:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.392551 5117 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.486826 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.486866 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.486875 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.486890 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.486902 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:53Z","lastTransitionTime":"2026-03-21T00:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.492823 5117 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.590088 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.590173 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.590193 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.590309 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.590335 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:53Z","lastTransitionTime":"2026-03-21T00:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.593391 5117 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-etcd/etcd-crc" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.692565 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.692661 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.692684 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.693036 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.693303 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:53Z","lastTransitionTime":"2026-03-21T00:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.697456 5117 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.797360 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.797425 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.797445 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.797476 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.797499 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:53Z","lastTransitionTime":"2026-03-21T00:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.900818 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.900969 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.900984 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.901499 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:53 crc kubenswrapper[5117]: I0321 00:12:53.901531 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:53Z","lastTransitionTime":"2026-03-21T00:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.004582 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.004947 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.005086 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.005240 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.005423 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.108750 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.108823 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.108843 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.108869 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.108888 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.179910 5117 apiserver.go:52] "Watching apiserver" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.190502 5117 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.191348 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-qp2t9","openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5","openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6","openshift-network-diagnostics/network-check-target-fhkjl","openshift-network-operator/iptables-alerter-5jnd7","openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv","openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq","openshift-ovn-kubernetes/ovnkube-node-9kggs","openshift-dns/node-resolver-x6fv6","openshift-multus/multus-72hnj","openshift-network-node-identity/network-node-identity-dgvkt","openshift-image-registry/node-ca-4sk9p","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-multus/multus-additional-cni-plugins-gzrb6","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-machine-config-operator/machine-config-daemon-w9wqq"] Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.193007 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.193665 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.193878 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.194302 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.194377 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.194590 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.200958 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"metrics-tls\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.201907 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.207562 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.207716 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"kube-root-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.207769 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"env-overrides\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.208036 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"ovnkube-identity-cm\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.208216 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.209173 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-node-identity\"/\"network-node-identity-cert\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.209569 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"openshift-service-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.211249 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.213540 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.213600 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.213619 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.213646 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.213668 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.214060 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.214109 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.214239 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.218495 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.218587 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-tk7bt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.220783 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.221288 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.223958 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.225038 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.225932 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.225980 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.226286 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-l2v2m\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.226317 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.227608 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.228476 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.228615 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qp2t9" podUID="482a0770-6dc5-4682-bf1e-e3c085129f33" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.233999 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.242003 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.245877 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.246114 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.245936 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-g6kgg\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.246397 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.245908 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.252494 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.269788 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.273066 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.275934 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-nl8tp\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.276180 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.280170 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.284982 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.285205 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.285332 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-tjs74\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.285425 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.288571 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.307218 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.315889 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.315946 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.315960 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.315982 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.316000 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.316734 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.319031 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.320500 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-nwglk\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.321107 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.322740 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.336240 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc4541ce-7789-4670-bc75-5c2868e52ce0-webhook-cert\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.336387 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42dc6\" (UniqueName: \"kubernetes.io/projected/482a0770-6dc5-4682-bf1e-e3c085129f33-kube-api-access-42dc6\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.336460 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-var-lib-cni-bin\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.336505 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-run-multus-certs\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.336584 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-iptables-alerter-script\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.336806 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-systemd-units\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.336905 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a700740-bb4e-41e9-a9b3-516fc4546c5a-tmp-dir\") pod \"node-resolver-x6fv6\" (UID: \"3a700740-bb4e-41e9-a9b3-516fc4546c5a\") " pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.336979 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dsgwk\" (UniqueName: \"kubernetes.io/projected/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-kube-api-access-dsgwk\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337020 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bdb18fcd-7135-4474-ada3-a29788223262-serviceca\") pod \"node-ca-4sk9p\" (UID: \"bdb18fcd-7135-4474-ada3-a29788223262\") " pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337061 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-m7xz2\" (UniqueName: \"kubernetes.io/projected/34177974-8d82-49d2-a763-391d0df3bbd8-kube-api-access-m7xz2\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337101 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-netns\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337302 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-node-log\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337340 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-os-release\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337370 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-cni-binary-copy\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337403 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-socket-dir-parent\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337435 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-daemon-config\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337466 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-kubelet\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337497 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-etc-openvswitch\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337527 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-script-lib\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337558 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-run-netns\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337661 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-conf-dir\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337761 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-log-socket\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337826 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmjsx\" (UniqueName: \"kubernetes.io/projected/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-kube-api-access-mmjsx\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.337926 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-cni-dir\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338093 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-ovnkube-identity-cm\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338183 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-var-lib-openvswitch\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338224 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-config\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338285 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcpq2\" (UniqueName: \"kubernetes.io/projected/1a068441-be86-4149-89e1-e1bb6bb0569b-kube-api-access-lcpq2\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338320 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-host-slash\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338357 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338389 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-system-cni-dir\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338420 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-var-lib-kubelet\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338454 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-hostroot\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338491 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-openvswitch\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338522 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-netd\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338556 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338590 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-env-overrides\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338630 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovn-node-metrics-cert\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338662 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3a700740-bb4e-41e9-a9b3-516fc4546c5a-hosts-file\") pod \"node-resolver-x6fv6\" (UID: \"3a700740-bb4e-41e9-a9b3-516fc4546c5a\") " pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338693 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2ktw\" (UniqueName: \"kubernetes.io/projected/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-kube-api-access-s2ktw\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338735 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/34177974-8d82-49d2-a763-391d0df3bbd8-host-etc-kube\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338774 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338811 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-systemd\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338848 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1a068441-be86-4149-89e1-e1bb6bb0569b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338920 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.338965 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-cnibin\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.339000 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/34177974-8d82-49d2-a763-391d0df3bbd8-metrics-tls\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.339038 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-ovn\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.339080 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-bin\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.339122 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-run-k8s-cni-cncf-io\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.339160 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.339547 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.343400 5117 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.345523 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-ovnkube-identity-cm\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.347293 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-iptables-alerter-script\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.348140 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bdb18fcd-7135-4474-ada3-a29788223262-host\") pod \"node-ca-4sk9p\" (UID: \"bdb18fcd-7135-4474-ada3-a29788223262\") " pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.356181 5117 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.356343 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-ovn-kubernetes\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.356662 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7k8x\" (UniqueName: \"kubernetes.io/projected/3a700740-bb4e-41e9-a9b3-516fc4546c5a-kube-api-access-j7k8x\") pod \"node-resolver-x6fv6\" (UID: \"3a700740-bb4e-41e9-a9b3-516fc4546c5a\") " pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.357449 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:54.857358606 +0000 UTC m=+88.150645808 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.359209 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc4541ce-7789-4670-bc75-5c2868e52ce0-webhook-cert\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.359736 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-env-overrides\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.360121 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8nt2j\" (UniqueName: \"kubernetes.io/projected/fc4541ce-7789-4670-bc75-5c2868e52ce0-kube-api-access-8nt2j\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.359963 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-x6fv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a700740-bb4e-41e9-a9b3-516fc4546c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7k8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x6fv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.360413 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-env-overrides\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.360639 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-env-overrides\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.360672 5117 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.360947 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.361090 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-var-lib-cni-multus\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.361274 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:54.86105928 +0000 UTC m=+88.154346492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.361368 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-etc-kubernetes\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.361411 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-slash\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.361734 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-ovnkube-config\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.361798 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nlcs\" (UniqueName: \"kubernetes.io/projected/bdb18fcd-7135-4474-ada3-a29788223262-kube-api-access-2nlcs\") pod \"node-ca-4sk9p\" (UID: \"bdb18fcd-7135-4474-ada3-a29788223262\") " pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.362748 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/34177974-8d82-49d2-a763-391d0df3bbd8-metrics-tls\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.366667 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7xz2\" (UniqueName: \"kubernetes.io/projected/34177974-8d82-49d2-a763-391d0df3bbd8-kube-api-access-m7xz2\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.369283 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.369315 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.369341 5117 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.369442 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:54.869419815 +0000 UTC m=+88.162707017 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.372642 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsgwk\" (UniqueName: \"kubernetes.io/projected/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-kube-api-access-dsgwk\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.386700 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.386872 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.387024 5117 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.387241 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:54.887209328 +0000 UTC m=+88.180496540 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.391711 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nt2j\" (UniqueName: \"kubernetes.io/projected/fc4541ce-7789-4670-bc75-5c2868e52ce0-kube-api-access-8nt2j\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.392318 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9kggs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.400117 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.403330 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qp2t9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482a0770-6dc5-4682-bf1e-e3c085129f33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42dc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42dc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qp2t9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.403392 5117 scope.go:117] "RemoveContainer" containerID="79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.404523 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\"" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.404839 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.405487 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.405630 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.405829 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"proxy-tls\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.406001 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-w9nzh\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.418306 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.418365 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.418376 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.418393 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.418405 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.419148 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d4e366f-7f9a-45ea-9ccd-505c06a20f24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://94684ef43a4caf326c026a0b8b43fe90d3f86c9c60562f24fa305fcfc6f70788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://ad24a3d986e06d5c3f1905aa7e44d8fcc888cb7acc101ca3cbb39adac8684500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://d9eee0d1d08899481bf9b5b0fe5429a679042a29b12e717e65d1a11a9ab2d79e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ad515cb5c17c39877403cc2a002a95cf96b2d32f46c7cb4b753d390ad268c3dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.430325 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7b9c1f1-3a7c-4cad-9c70-09dc9f7e75b5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c2514a311c8df4b1faca8a81fa4c3becfae5f7e05e32fe441cc7d61e4b3fbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://000182f0824f8c8f7487c695f547ec126efb5f50836571807732678886484fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://000182f0824f8c8f7487c695f547ec126efb5f50836571807732678886484fe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.452731 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eae805bb-c929-4ab2-af7b-3e582255fe6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"},\\\"containerID\\\":\\\"cri-o://f6e2280de7b69beabb5b164fe42e7440579b41b13e07f84fcd7a6ad773218de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://825d8b369cbcdc65a64926c868202d0b3b9253162f40994fa0f447a5beeede1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ed17e61de6c00e7e405bc7beaa0ce1f7e26b6a4ec68e83ebee1f4c7ec0d3f27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b743995a17c55f75fba00e6815f49b6fc598d27f4c4c81452bad9ca5806d9f80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://b44d38825bd92498e91533afc0db62985336c2b9293ccdc5221262dcacd0acde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:31Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://1f7effd8d591e2ac11bcdb21df51a7b48ab90a589f431a721326bea14181aa6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7effd8d591e2ac11bcdb21df51a7b48ab90a589f431a721326bea14181aa6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd-auto-backup\\\",\\\"name\\\":\\\"etcd-auto-backup-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://c2f608613c2ee33e9d147e0481d3f47512c4db2c0e5fe416c84ad073975f3d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f608613c2ee33e9d147e0481d3f47512c4db2c0e5fe416c84ad073975f3d5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://d6e7a9fe3510c2f332f71cb736921b49b751c42424d2cbb5f6e9cb6c3726555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e7a9fe3510c2f332f71cb736921b49b751c42424d2cbb5f6e9cb6c3726555e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.462193 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3a700740-bb4e-41e9-a9b3-516fc4546c5a-hosts-file\") pod \"node-resolver-x6fv6\" (UID: \"3a700740-bb4e-41e9-a9b3-516fc4546c5a\") " pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.462556 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-s2ktw\" (UniqueName: \"kubernetes.io/projected/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-kube-api-access-s2ktw\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.462753 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e685d70e-7b85-464d-b251-f682148f431f-cni-binary-copy\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.462932 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/34177974-8d82-49d2-a763-391d0df3bbd8-host-etc-kube\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.462303 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3a700740-bb4e-41e9-a9b3-516fc4546c5a-hosts-file\") pod \"node-resolver-x6fv6\" (UID: \"3a700740-bb4e-41e9-a9b3-516fc4546c5a\") " pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463024 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/34177974-8d82-49d2-a763-391d0df3bbd8-host-etc-kube\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463136 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-systemd\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463243 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1a068441-be86-4149-89e1-e1bb6bb0569b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463330 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-cnibin\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463376 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-ovn\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463400 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-bin\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463420 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-run-k8s-cni-cncf-io\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463474 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bdb18fcd-7135-4474-ada3-a29788223262-host\") pod \"node-ca-4sk9p\" (UID: \"bdb18fcd-7135-4474-ada3-a29788223262\") " pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463516 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e685d70e-7b85-464d-b251-f682148f431f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463569 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-ovn-kubernetes\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463610 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-j7k8x\" (UniqueName: \"kubernetes.io/projected/3a700740-bb4e-41e9-a9b3-516fc4546c5a-kube-api-access-j7k8x\") pod \"node-resolver-x6fv6\" (UID: \"3a700740-bb4e-41e9-a9b3-516fc4546c5a\") " pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463640 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-env-overrides\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463691 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-var-lib-cni-multus\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463692 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-cnibin\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463769 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-ovn-kubernetes\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463774 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-etc-kubernetes\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463716 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-etc-kubernetes\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463798 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-bin\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463860 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-slash\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463836 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-slash\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463904 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-var-lib-cni-multus\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463938 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-ovnkube-config\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463993 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2nlcs\" (UniqueName: \"kubernetes.io/projected/bdb18fcd-7135-4474-ada3-a29788223262-kube-api-access-2nlcs\") pod \"node-ca-4sk9p\" (UID: \"bdb18fcd-7135-4474-ada3-a29788223262\") " pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.463675 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-ovn\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464084 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-run-k8s-cni-cncf-io\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464084 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bdb18fcd-7135-4474-ada3-a29788223262-host\") pod \"node-ca-4sk9p\" (UID: \"bdb18fcd-7135-4474-ada3-a29788223262\") " pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464347 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-os-release\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464430 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-42dc6\" (UniqueName: \"kubernetes.io/projected/482a0770-6dc5-4682-bf1e-e3c085129f33-kube-api-access-42dc6\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464459 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-var-lib-cni-bin\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464481 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-run-multus-certs\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464524 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-rootfs\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464546 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-systemd-units\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464568 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a700740-bb4e-41e9-a9b3-516fc4546c5a-tmp-dir\") pod \"node-resolver-x6fv6\" (UID: \"3a700740-bb4e-41e9-a9b3-516fc4546c5a\") " pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464604 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bdb18fcd-7135-4474-ada3-a29788223262-serviceca\") pod \"node-ca-4sk9p\" (UID: \"bdb18fcd-7135-4474-ada3-a29788223262\") " pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464626 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/e685d70e-7b85-464d-b251-f682148f431f-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464657 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-netns\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464674 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-node-log\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464691 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-env-overrides\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464724 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-os-release\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464775 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-var-lib-cni-bin\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464781 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-systemd-units\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464808 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-node-log\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464796 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-netns\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.464851 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-run-multus-certs\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465263 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a700740-bb4e-41e9-a9b3-516fc4546c5a-tmp-dir\") pod \"node-resolver-x6fv6\" (UID: \"3a700740-bb4e-41e9-a9b3-516fc4546c5a\") " pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465420 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-cni-binary-copy\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465467 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-socket-dir-parent\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465526 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-daemon-config\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465549 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-kubelet\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465565 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-etc-openvswitch\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465602 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-script-lib\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465650 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-run-netns\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465690 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-conf-dir\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465716 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-cnibin\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465750 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-os-release\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465773 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-log-socket\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465785 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-kubelet\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465816 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-log-socket\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465828 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-ovnkube-config\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465871 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-etc-openvswitch\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465907 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-run-netns\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465927 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-conf-dir\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.465943 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bdb18fcd-7135-4474-ada3-a29788223262-serviceca\") pod \"node-ca-4sk9p\" (UID: \"bdb18fcd-7135-4474-ada3-a29788223262\") " pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466012 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-socket-dir-parent\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466039 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mmjsx\" (UniqueName: \"kubernetes.io/projected/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-kube-api-access-mmjsx\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466106 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-cni-dir\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466166 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-mcd-auth-proxy-config\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466194 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jq85\" (UniqueName: \"kubernetes.io/projected/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-kube-api-access-6jq85\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466261 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-var-lib-openvswitch\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466286 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-cni-dir\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466294 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-config\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466346 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-script-lib\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.466967 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lcpq2\" (UniqueName: \"kubernetes.io/projected/1a068441-be86-4149-89e1-e1bb6bb0569b-kube-api-access-lcpq2\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.467047 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-host-slash\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.467134 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.467162 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-config\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.467184 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-system-cni-dir\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.467233 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-var-lib-kubelet\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.467597 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1a068441-be86-4149-89e1-e1bb6bb0569b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.467996 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-hostroot\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468072 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-host-var-lib-kubelet\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468135 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-host-slash\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468163 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-var-lib-openvswitch\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.468409 5117 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468417 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-proxy-tls\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468464 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.468479 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs podName:482a0770-6dc5-4682-bf1e-e3c085129f33 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:54.96846254 +0000 UTC m=+88.261749712 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs") pod "network-metrics-daemon-qp2t9" (UID: "482a0770-6dc5-4682-bf1e-e3c085129f33") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468576 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-system-cni-dir\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468658 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-hostroot\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468764 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-openvswitch\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468865 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-netd\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.468945 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44hwn\" (UniqueName: \"kubernetes.io/projected/e685d70e-7b85-464d-b251-f682148f431f-kube-api-access-44hwn\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.469042 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-openvswitch\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.469068 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.469100 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-netd\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.469145 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-env-overrides\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.469161 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.469126 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.469280 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-system-cni-dir\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.469403 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovn-node-metrics-cert\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.469765 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-env-overrides\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.470119 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-systemd\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.470184 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-cni-binary-copy\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.470636 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-multus-daemon-config\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.475290 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovn-node-metrics-cert\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.481092 5117 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.482806 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7k8x\" (UniqueName: \"kubernetes.io/projected/3a700740-bb4e-41e9-a9b3-516fc4546c5a-kube-api-access-j7k8x\") pod \"node-resolver-x6fv6\" (UID: \"3a700740-bb4e-41e9-a9b3-516fc4546c5a\") " pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.485361 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2ktw\" (UniqueName: \"kubernetes.io/projected/c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e-kube-api-access-s2ktw\") pod \"multus-72hnj\" (UID: \"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\") " pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.485755 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nlcs\" (UniqueName: \"kubernetes.io/projected/bdb18fcd-7135-4474-ada3-a29788223262-kube-api-access-2nlcs\") pod \"node-ca-4sk9p\" (UID: \"bdb18fcd-7135-4474-ada3-a29788223262\") " pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.487715 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmjsx\" (UniqueName: \"kubernetes.io/projected/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-kube-api-access-mmjsx\") pod \"ovnkube-node-9kggs\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.488659 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.491866 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcpq2\" (UniqueName: \"kubernetes.io/projected/1a068441-be86-4149-89e1-e1bb6bb0569b-kube-api-access-lcpq2\") pod \"ovnkube-control-plane-57b78d8988-f4xkq\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.494727 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-42dc6\" (UniqueName: \"kubernetes.io/projected/482a0770-6dc5-4682-bf1e-e3c085129f33-kube-api-access-42dc6\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.499351 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.516703 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-72hnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ktw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-72hnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.520671 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.520701 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.520710 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.520725 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.520735 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.529852 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qp2t9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482a0770-6dc5-4682-bf1e-e3c085129f33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42dc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42dc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qp2t9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.538138 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4sk9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdb18fcd-7135-4474-ada3-a29788223262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nlcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4sk9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.541374 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.548767 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jq85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jq85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9wqq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.559337 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-dgvkt" Mar 21 00:12:54 crc kubenswrapper[5117]: W0321 00:12:54.559676 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34177974_8d82_49d2_a763_391d0df3bbd8.slice/crio-969e55eb73049032c3b8f49bcecae7efa8dc9fa359e490a19d6cafebcea05fef WatchSource:0}: Error finding container 969e55eb73049032c3b8f49bcecae7efa8dc9fa359e490a19d6cafebcea05fef: Status 404 returned error can't find the container with id 969e55eb73049032c3b8f49bcecae7efa8dc9fa359e490a19d6cafebcea05fef Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.560819 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e2c803-23f7-446f-a7f3-90de8dc4355c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://af1c6802eebfabdf424eae7ad9eecfcaaffcd44a0d2cfad7300f2ff7a532cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://3a5498baab91a2da6643fcdc8b35141226196ffe302f6c2f38a266e518799587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://cb3b6dc02877e4af1eb1e9c2de8b9041ba203d71b6e52185287a6cd4c3ba28b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c6f7650b6653e951c6fe6e6883d205ee80691d8683967540c1766ef0d1df7f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6f7650b6653e951c6fe6e6883d205ee80691d8683967540c1766ef0d1df7f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570286 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570331 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570373 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570399 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570427 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgrkj\" (UniqueName: \"kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj\") pod \"42a11a02-47e1-488f-b270-2679d3298b0e\" (UID: \"42a11a02-47e1-488f-b270-2679d3298b0e\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570452 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570476 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570506 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570536 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570562 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs\") pod \"593a3561-7760-45c5-8f91-5aaef7475d0f\" (UID: \"593a3561-7760-45c5-8f91-5aaef7475d0f\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570587 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570612 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle\") pod \"ce090a97-9ab6-4c40-a719-64ff2acd9778\" (UID: \"ce090a97-9ab6-4c40-a719-64ff2acd9778\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570638 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g8ts\" (UniqueName: \"kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570668 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570692 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dztfv\" (UniqueName: \"kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570719 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570750 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pddnv\" (UniqueName: \"kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv\") pod \"e093be35-bb62-4843-b2e8-094545761610\" (UID: \"e093be35-bb62-4843-b2e8-094545761610\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570781 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vsz9\" (UniqueName: \"kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570816 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570845 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570875 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570902 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m26jq\" (UniqueName: \"kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570931 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570966 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.570992 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571021 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571053 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571082 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca\") pod \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\" (UID: \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571109 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks6v2\" (UniqueName: \"kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2\") pod \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\" (UID: \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571138 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfzkj\" (UniqueName: \"kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj\") pod \"0effdbcf-dd7d-404d-9d48-77536d665a5d\" (UID: \"0effdbcf-dd7d-404d-9d48-77536d665a5d\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571162 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571186 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkdh6\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571212 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571241 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571282 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571308 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxfcv\" (UniqueName: \"kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv\") pod \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\" (UID: \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.571375 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.572237 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.572495 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv" (OuterVolumeSpecName: "kube-api-access-pddnv") pod "e093be35-bb62-4843-b2e8-094545761610" (UID: "e093be35-bb62-4843-b2e8-094545761610"). InnerVolumeSpecName "kube-api-access-pddnv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.572838 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.572991 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.573101 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv" (OuterVolumeSpecName: "kube-api-access-dztfv") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "kube-api-access-dztfv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.573413 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.573508 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.573566 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.573775 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts" (OuterVolumeSpecName: "kube-api-access-4g8ts") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "kube-api-access-4g8ts". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.573784 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj" (OuterVolumeSpecName: "kube-api-access-qgrkj") pod "42a11a02-47e1-488f-b270-2679d3298b0e" (UID: "42a11a02-47e1-488f-b270-2679d3298b0e"). InnerVolumeSpecName "kube-api-access-qgrkj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.573905 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "ce090a97-9ab6-4c40-a719-64ff2acd9778" (UID: "ce090a97-9ab6-4c40-a719-64ff2acd9778"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574371 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq" (OuterVolumeSpecName: "kube-api-access-m26jq") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "kube-api-access-m26jq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574451 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities\") pod \"149b3c48-e17c-4a66-a835-d86dabf6ff13\" (UID: \"149b3c48-e17c-4a66-a835-d86dabf6ff13\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574502 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9" (OuterVolumeSpecName: "kube-api-access-9vsz9") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "kube-api-access-9vsz9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574507 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp" (OuterVolumeSpecName: "tmp") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574505 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities\") pod \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\" (UID: \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574557 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574580 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp" (OuterVolumeSpecName: "tmp") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574586 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574624 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token\") pod \"593a3561-7760-45c5-8f91-5aaef7475d0f\" (UID: \"593a3561-7760-45c5-8f91-5aaef7475d0f\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574649 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs\") pod \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\" (UID: \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574668 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content\") pod \"94a6e063-3d1a-4d44-875d-185291448c31\" (UID: \"94a6e063-3d1a-4d44-875d-185291448c31\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574690 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config\") pod \"7599e0b6-bddf-4def-b7f2-0b32206e8651\" (UID: \"7599e0b6-bddf-4def-b7f2-0b32206e8651\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574821 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574844 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities\") pod \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\" (UID: \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574863 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574880 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574898 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs\") pod \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\" (UID: \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574915 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574937 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9stx\" (UniqueName: \"kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574960 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574985 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575087 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575119 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575144 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls\") pod \"6077b63e-53a2-4f96-9d56-1ce0324e4913\" (UID: \"6077b63e-53a2-4f96-9d56-1ce0324e4913\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575170 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575197 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575221 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575239 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfp5s\" (UniqueName: \"kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s\") pod \"cc85e424-18b2-4924-920b-bd291a8c4b01\" (UID: \"cc85e424-18b2-4924-920b-bd291a8c4b01\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575306 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575332 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575357 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key\") pod \"ce090a97-9ab6-4c40-a719-64ff2acd9778\" (UID: \"ce090a97-9ab6-4c40-a719-64ff2acd9778\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575374 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls\") pod \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\" (UID: \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575392 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575412 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576311 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nspp\" (UniqueName: \"kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576347 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576429 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576453 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z4sw\" (UniqueName: \"kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw\") pod \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\" (UID: \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576472 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576501 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7cps\" (UniqueName: \"kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps\") pod \"af41de71-79cf-4590-bbe9-9e8b848862cb\" (UID: \"af41de71-79cf-4590-bbe9-9e8b848862cb\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576526 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576548 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576571 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576600 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hckvg\" (UniqueName: \"kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576628 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4smf\" (UniqueName: \"kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf\") pod \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\" (UID: \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576658 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities\") pod \"cc85e424-18b2-4924-920b-bd291a8c4b01\" (UID: \"cc85e424-18b2-4924-920b-bd291a8c4b01\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576678 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576709 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content\") pod \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\" (UID: \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576728 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576748 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgx6b\" (UniqueName: \"kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b\") pod \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\" (UID: \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576768 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576788 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576807 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls\") pod \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\" (UID: \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576826 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576846 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsb9b\" (UniqueName: \"kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b\") pod \"09cfa50b-4138-4585-a53e-64dd3ab73335\" (UID: \"09cfa50b-4138-4585-a53e-64dd3ab73335\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576864 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content\") pod \"149b3c48-e17c-4a66-a835-d86dabf6ff13\" (UID: \"149b3c48-e17c-4a66-a835-d86dabf6ff13\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576885 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grwfz\" (UniqueName: \"kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz\") pod \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\" (UID: \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576910 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576938 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content\") pod \"b605f283-6f2e-42da-a838-54421690f7d0\" (UID: \"b605f283-6f2e-42da-a838-54421690f7d0\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576957 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.578609 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.578647 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.578673 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.578699 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580101 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content\") pod \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\" (UID: \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580170 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580227 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert\") pod \"d7e8f42f-dc0e-424b-bb56-5ec849834888\" (UID: \"d7e8f42f-dc0e-424b-bb56-5ec849834888\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580260 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580281 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580306 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580326 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580345 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pskd\" (UniqueName: \"kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580365 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580385 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5rsr\" (UniqueName: \"kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580407 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir\") pod \"6077b63e-53a2-4f96-9d56-1ce0324e4913\" (UID: \"6077b63e-53a2-4f96-9d56-1ce0324e4913\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580425 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftwb6\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580442 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hb7m\" (UniqueName: \"kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m\") pod \"94a6e063-3d1a-4d44-875d-185291448c31\" (UID: \"94a6e063-3d1a-4d44-875d-185291448c31\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580460 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5lgh\" (UniqueName: \"kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580481 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580499 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g4lr\" (UniqueName: \"kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr\") pod \"f7e2c886-118e-43bb-bef1-c78134de392b\" (UID: \"f7e2c886-118e-43bb-bef1-c78134de392b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.574782 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6" (OuterVolumeSpecName: "kube-api-access-tkdh6") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "kube-api-access-tkdh6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575173 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities" (OuterVolumeSpecName: "utilities") pod "584e1f4a-8205-47d7-8efb-3afc6017c4c9" (UID: "584e1f4a-8205-47d7-8efb-3afc6017c4c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.575323 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config" (OuterVolumeSpecName: "config") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576419 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576734 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.576839 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "593a3561-7760-45c5-8f91-5aaef7475d0f" (UID: "593a3561-7760-45c5-8f91-5aaef7475d0f"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: W0321 00:12:54.576903 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc4541ce_7789_4670_bc75_5c2868e52ce0.slice/crio-f267fac5a93802824762b8a2a59b78bb435633b8df58049bc07de79086f53bcb WatchSource:0}: Error finding container f267fac5a93802824762b8a2a59b78bb435633b8df58049bc07de79086f53bcb: Status 404 returned error can't find the container with id f267fac5a93802824762b8a2a59b78bb435633b8df58049bc07de79086f53bcb Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.577712 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.577910 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.578485 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.578518 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.578639 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2" (OuterVolumeSpecName: "kube-api-access-ks6v2") pod "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" (UID: "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a"). InnerVolumeSpecName "kube-api-access-ks6v2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.579151 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv" (OuterVolumeSpecName: "kube-api-access-xxfcv") pod "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" (UID: "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff"). InnerVolumeSpecName "kube-api-access-xxfcv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.579386 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "e1d2a42d-af1d-4054-9618-ab545e0ed8b7" (UID: "e1d2a42d-af1d-4054-9618-ab545e0ed8b7"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.579369 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.579456 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.579598 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps" (OuterVolumeSpecName: "kube-api-access-d7cps") pod "af41de71-79cf-4590-bbe9-9e8b848862cb" (UID: "af41de71-79cf-4590-bbe9-9e8b848862cb"). InnerVolumeSpecName "kube-api-access-d7cps". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.579630 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities" (OuterVolumeSpecName: "utilities") pod "cc85e424-18b2-4924-920b-bd291a8c4b01" (UID: "cc85e424-18b2-4924-920b-bd291a8c4b01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.579951 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs" (OuterVolumeSpecName: "certs") pod "593a3561-7760-45c5-8f91-5aaef7475d0f" (UID: "593a3561-7760-45c5-8f91-5aaef7475d0f"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580071 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw" (OuterVolumeSpecName: "kube-api-access-9z4sw") pod "e1d2a42d-af1d-4054-9618-ab545e0ed8b7" (UID: "e1d2a42d-af1d-4054-9618-ab545e0ed8b7"). InnerVolumeSpecName "kube-api-access-9z4sw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580133 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580243 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580462 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.581665 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b605f283-6f2e-42da-a838-54421690f7d0" (UID: "b605f283-6f2e-42da-a838-54421690f7d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.581686 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr" (OuterVolumeSpecName: "kube-api-access-6g4lr") pod "f7e2c886-118e-43bb-bef1-c78134de392b" (UID: "f7e2c886-118e-43bb-bef1-c78134de392b"). InnerVolumeSpecName "kube-api-access-6g4lr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.581892 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj" (OuterVolumeSpecName: "kube-api-access-mfzkj") pod "0effdbcf-dd7d-404d-9d48-77536d665a5d" (UID: "0effdbcf-dd7d-404d-9d48-77536d665a5d"). InnerVolumeSpecName "kube-api-access-mfzkj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.581984 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "0dd0fbac-8c0d-4228-8faa-abbeedabf7db" (UID: "0dd0fbac-8c0d-4228-8faa-abbeedabf7db"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.582024 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b" (OuterVolumeSpecName: "kube-api-access-pgx6b") pod "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" (UID: "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4"). InnerVolumeSpecName "kube-api-access-pgx6b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.582053 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.582319 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.582321 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config" (OuterVolumeSpecName: "config") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.582401 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config" (OuterVolumeSpecName: "config") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.582498 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config" (OuterVolumeSpecName: "config") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.582588 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b" (OuterVolumeSpecName: "kube-api-access-zsb9b") pod "09cfa50b-4138-4585-a53e-64dd3ab73335" (UID: "09cfa50b-4138-4585-a53e-64dd3ab73335"). InnerVolumeSpecName "kube-api-access-zsb9b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580658 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580680 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580743 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca" (OuterVolumeSpecName: "serviceca") pod "5ebfebf6-3ecd-458e-943f-bb25b52e2718" (UID: "5ebfebf6-3ecd-458e-943f-bb25b52e2718"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.580947 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.581067 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.581355 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:12:55.081330108 +0000 UTC m=+88.374617280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.582973 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583001 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities\") pod \"b605f283-6f2e-42da-a838-54421690f7d0\" (UID: \"b605f283-6f2e-42da-a838-54421690f7d0\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583033 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583056 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm9x7\" (UniqueName: \"kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583071 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf" (OuterVolumeSpecName: "kube-api-access-q4smf") pod "0dd0fbac-8c0d-4228-8faa-abbeedabf7db" (UID: "0dd0fbac-8c0d-4228-8faa-abbeedabf7db"). InnerVolumeSpecName "kube-api-access-q4smf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583078 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583125 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptkcf\" (UniqueName: \"kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf\") pod \"7599e0b6-bddf-4def-b7f2-0b32206e8651\" (UID: \"7599e0b6-bddf-4def-b7f2-0b32206e8651\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583146 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26xrl\" (UniqueName: \"kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl\") pod \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\" (UID: \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583165 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583183 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rmnv\" (UniqueName: \"kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv\") pod \"b605f283-6f2e-42da-a838-54421690f7d0\" (UID: \"b605f283-6f2e-42da-a838-54421690f7d0\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583205 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls\") pod \"42a11a02-47e1-488f-b270-2679d3298b0e\" (UID: \"42a11a02-47e1-488f-b270-2679d3298b0e\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583225 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583244 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities\") pod \"94a6e063-3d1a-4d44-875d-185291448c31\" (UID: \"94a6e063-3d1a-4d44-875d-185291448c31\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583286 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqbfk\" (UniqueName: \"kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583309 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583331 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583350 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583369 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583389 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content\") pod \"cc85e424-18b2-4924-920b-bd291a8c4b01\" (UID: \"cc85e424-18b2-4924-920b-bd291a8c4b01\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583391 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd" (OuterVolumeSpecName: "kube-api-access-8pskd") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "kube-api-access-8pskd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583408 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert\") pod \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\" (UID: \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583428 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config\") pod \"09cfa50b-4138-4585-a53e-64dd3ab73335\" (UID: \"09cfa50b-4138-4585-a53e-64dd3ab73335\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583449 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583470 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content\") pod \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\" (UID: \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583490 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twvbl\" (UniqueName: \"kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl\") pod \"b4750666-1362-4001-abd0-6f89964cc621\" (UID: \"b4750666-1362-4001-abd0-6f89964cc621\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583508 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content\") pod \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\" (UID: \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583530 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583510 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583556 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583672 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583761 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583805 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert\") pod \"01080b46-74f1-4191-8755-5152a57b3b25\" (UID: \"01080b46-74f1-4191-8755-5152a57b3b25\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583849 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert\") pod \"09cfa50b-4138-4585-a53e-64dd3ab73335\" (UID: \"09cfa50b-4138-4585-a53e-64dd3ab73335\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583890 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config\") pod \"81e39f7b-62e4-4fc9-992a-6535ce127a02\" (UID: \"81e39f7b-62e4-4fc9-992a-6535ce127a02\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583934 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.583972 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584017 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584073 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "584e1f4a-8205-47d7-8efb-3afc6017c4c9" (UID: "584e1f4a-8205-47d7-8efb-3afc6017c4c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584108 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584164 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584202 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert\") pod \"16bdd140-dce1-464c-ab47-dd5798d1d256\" (UID: \"16bdd140-dce1-464c-ab47-dd5798d1d256\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584296 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities\") pod \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\" (UID: \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584364 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg8nc\" (UniqueName: \"kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584514 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbc2l\" (UniqueName: \"kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l\") pod \"593a3561-7760-45c5-8f91-5aaef7475d0f\" (UID: \"593a3561-7760-45c5-8f91-5aaef7475d0f\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584558 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584645 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddlk9\" (UniqueName: \"kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584671 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.584734 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585037 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca" (OuterVolumeSpecName: "client-ca") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585218 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585274 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585305 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585328 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585356 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585381 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585403 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585425 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585446 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585455 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg" (OuterVolumeSpecName: "kube-api-access-hckvg") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "kube-api-access-hckvg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585469 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access\") pod \"d7e8f42f-dc0e-424b-bb56-5ec849834888\" (UID: \"d7e8f42f-dc0e-424b-bb56-5ec849834888\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585491 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dmhf\" (UniqueName: \"kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585513 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585535 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99zj9\" (UniqueName: \"kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585506 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp" (OuterVolumeSpecName: "kube-api-access-8nspp") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "kube-api-access-8nspp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585554 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585573 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnxbn\" (UniqueName: \"kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn\") pod \"ce090a97-9ab6-4c40-a719-64ff2acd9778\" (UID: \"ce090a97-9ab6-4c40-a719-64ff2acd9778\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585600 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585620 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config\") pod \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\" (UID: \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585619 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz" (OuterVolumeSpecName: "kube-api-access-grwfz") pod "31fa8943-81cc-4750-a0b7-0fa9ab5af883" (UID: "31fa8943-81cc-4750-a0b7-0fa9ab5af883"). InnerVolumeSpecName "kube-api-access-grwfz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585641 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585736 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jjkz\" (UniqueName: \"kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585775 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585799 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585839 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585860 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pllx6\" (UniqueName: \"kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6\") pod \"81e39f7b-62e4-4fc9-992a-6535ce127a02\" (UID: \"81e39f7b-62e4-4fc9-992a-6535ce127a02\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585841 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf" (OuterVolumeSpecName: "kube-api-access-ptkcf") pod "7599e0b6-bddf-4def-b7f2-0b32206e8651" (UID: "7599e0b6-bddf-4def-b7f2-0b32206e8651"). InnerVolumeSpecName "kube-api-access-ptkcf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585888 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.585953 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586008 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586331 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586408 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities\") pod \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\" (UID: \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586460 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config\") pod \"01080b46-74f1-4191-8755-5152a57b3b25\" (UID: \"01080b46-74f1-4191-8755-5152a57b3b25\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586515 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586581 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4tqq\" (UniqueName: \"kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq\") pod \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\" (UID: \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586628 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94l9h\" (UniqueName: \"kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h\") pod \"16bdd140-dce1-464c-ab47-dd5798d1d256\" (UID: \"16bdd140-dce1-464c-ab47-dd5798d1d256\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587095 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.590193 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-5jnd7" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.590204 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.591193 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.591224 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzt4w\" (UniqueName: \"kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w\") pod \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\" (UID: \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.591293 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.591349 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmmzf\" (UniqueName: \"kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.591377 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zth6t\" (UniqueName: \"kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t\") pod \"6077b63e-53a2-4f96-9d56-1ce0324e4913\" (UID: \"6077b63e-53a2-4f96-9d56-1ce0324e4913\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.591275 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797d5d7c-c536-4487-99be-f1f0219b63c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T00:12:42Z\\\",\\\"message\\\":\\\"vvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nW0321 00:12:42.499742 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 00:12:42.499925 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI0321 00:12:42.501331 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3355725928/tls.crt::/tmp/serving-cert-3355725928/tls.key\\\\\\\"\\\\nI0321 00:12:42.934594 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 00:12:42.943210 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 00:12:42.943284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 00:12:42.943339 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 00:12:42.943351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 00:12:42.951353 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 00:12:42.951415 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 00:12:42.951428 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 00:12:42.951440 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0321 00:12:42.951448 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 00:12:42.951457 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 00:12:42.951465 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0321 00:12:42.951439 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0321 00:12:42.954854 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T00:12:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592014 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nb9c\" (UniqueName: \"kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592051 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbmqg\" (UniqueName: \"kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592077 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592102 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592124 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592488 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tknt7\" (UniqueName: \"kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7\") pod \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\" (UID: \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592512 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592538 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592559 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config\") pod \"b4750666-1362-4001-abd0-6f89964cc621\" (UID: \"b4750666-1362-4001-abd0-6f89964cc621\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592582 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert\") pod \"7599e0b6-bddf-4def-b7f2-0b32206e8651\" (UID: \"7599e0b6-bddf-4def-b7f2-0b32206e8651\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592604 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592628 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l87hs\" (UniqueName: \"kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs\") pod \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\" (UID: \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592662 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws8zz\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592684 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592732 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates\") pod \"16bdd140-dce1-464c-ab47-dd5798d1d256\" (UID: \"16bdd140-dce1-464c-ab47-dd5798d1d256\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592755 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy\") pod \"81e39f7b-62e4-4fc9-992a-6535ce127a02\" (UID: \"81e39f7b-62e4-4fc9-992a-6535ce127a02\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592776 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592797 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj4qr\" (UniqueName: \"kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr\") pod \"149b3c48-e17c-4a66-a835-d86dabf6ff13\" (UID: \"149b3c48-e17c-4a66-a835-d86dabf6ff13\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592819 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lcfw\" (UniqueName: \"kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592840 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592862 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592884 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir\") pod \"f7e2c886-118e-43bb-bef1-c78134de392b\" (UID: \"f7e2c886-118e-43bb-bef1-c78134de392b\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592908 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls\") pod \"b4750666-1362-4001-abd0-6f89964cc621\" (UID: \"b4750666-1362-4001-abd0-6f89964cc621\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.593764 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca\") pod \"d7e8f42f-dc0e-424b-bb56-5ec849834888\" (UID: \"d7e8f42f-dc0e-424b-bb56-5ec849834888\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.593798 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.593844 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.593884 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.593924 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.593949 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjwtd\" (UniqueName: \"kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594048 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w94wk\" (UniqueName: \"kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk\") pod \"01080b46-74f1-4191-8755-5152a57b3b25\" (UID: \"01080b46-74f1-4191-8755-5152a57b3b25\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594100 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert\") pod \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\" (UID: \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\") " Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594226 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-cnibin\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594306 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-mcd-auth-proxy-config\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594329 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6jq85\" (UniqueName: \"kubernetes.io/projected/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-kube-api-access-6jq85\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594390 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-proxy-tls\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594411 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594443 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-44hwn\" (UniqueName: \"kubernetes.io/projected/e685d70e-7b85-464d-b251-f682148f431f-kube-api-access-44hwn\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594473 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-system-cni-dir\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594516 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e685d70e-7b85-464d-b251-f682148f431f-cni-binary-copy\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594598 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e685d70e-7b85-464d-b251-f682148f431f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594669 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-os-release\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594703 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-rootfs\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594734 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/e685d70e-7b85-464d-b251-f682148f431f-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594874 5117 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594889 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ptkcf\" (UniqueName: \"kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594903 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594914 5117 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594925 5117 reconciler_common.go:299] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594935 5117 reconciler_common.go:299] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594946 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qgrkj\" (UniqueName: \"kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594957 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594970 5117 reconciler_common.go:299] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594985 5117 reconciler_common.go:299] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.594998 5117 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595011 5117 reconciler_common.go:299] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595024 5117 reconciler_common.go:299] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595037 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4g8ts\" (UniqueName: \"kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595050 5117 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595061 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dztfv\" (UniqueName: \"kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595070 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595081 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pddnv\" (UniqueName: \"kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595091 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9vsz9\" (UniqueName: \"kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595104 5117 reconciler_common.go:299] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595123 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595132 5117 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595143 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-m26jq\" (UniqueName: \"kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595153 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595163 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595174 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595184 5117 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595195 5117 reconciler_common.go:299] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595205 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ks6v2\" (UniqueName: \"kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595216 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mfzkj\" (UniqueName: \"kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595225 5117 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595236 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tkdh6\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595260 5117 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595270 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595282 5117 reconciler_common.go:299] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595294 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xxfcv\" (UniqueName: \"kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595304 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595313 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595324 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595334 5117 reconciler_common.go:299] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595344 5117 reconciler_common.go:299] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595354 5117 reconciler_common.go:299] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595363 5117 reconciler_common.go:299] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595373 5117 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595383 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595393 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8nspp\" (UniqueName: \"kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595404 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9z4sw\" (UniqueName: \"kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595414 5117 reconciler_common.go:299] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595423 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d7cps\" (UniqueName: \"kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595433 5117 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595442 5117 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595452 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hckvg\" (UniqueName: \"kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595464 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-q4smf\" (UniqueName: \"kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595474 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595484 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595494 5117 reconciler_common.go:299] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595505 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pgx6b\" (UniqueName: \"kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595516 5117 reconciler_common.go:299] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595526 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zsb9b\" (UniqueName: \"kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595537 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-grwfz\" (UniqueName: \"kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595547 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595556 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595565 5117 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595575 5117 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595585 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8pskd\" (UniqueName: \"kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595595 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6g4lr\" (UniqueName: \"kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.602186 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/e685d70e-7b85-464d-b251-f682148f431f-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.604005 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-x6fv6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.606008 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-cnibin\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.606697 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-mcd-auth-proxy-config\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.613248 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.613467 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-os-release\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.613482 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e685d70e-7b85-464d-b251-f682148f431f-cni-binary-copy\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586332 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "6077b63e-53a2-4f96-9d56-1ce0324e4913" (UID: "6077b63e-53a2-4f96-9d56-1ce0324e4913"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.613697 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.613740 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.613967 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "a208c9c2-333b-4b4a-be0d-bc32ec38a821" (UID: "a208c9c2-333b-4b4a-be0d-bc32ec38a821"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.613992 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7599e0b6-bddf-4def-b7f2-0b32206e8651" (UID: "7599e0b6-bddf-4def-b7f2-0b32206e8651"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.614157 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586332 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config" (OuterVolumeSpecName: "console-config") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.586410 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587086 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca" (OuterVolumeSpecName: "service-ca") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587118 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key" (OuterVolumeSpecName: "signing-key") pod "ce090a97-9ab6-4c40-a719-64ff2acd9778" (UID: "ce090a97-9ab6-4c40-a719-64ff2acd9778"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.614358 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e685d70e-7b85-464d-b251-f682148f431f-system-cni-dir\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587270 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587290 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.614432 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e685d70e-7b85-464d-b251-f682148f431f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587340 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587532 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp" (OuterVolumeSpecName: "tmp") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587598 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587814 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "6077b63e-53a2-4f96-9d56-1ce0324e4913" (UID: "6077b63e-53a2-4f96-9d56-1ce0324e4913"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587843 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587915 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.587995 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s" (OuterVolumeSpecName: "kube-api-access-xfp5s") pod "cc85e424-18b2-4924-920b-bd291a8c4b01" (UID: "cc85e424-18b2-4924-920b-bd291a8c4b01"). InnerVolumeSpecName "kube-api-access-xfp5s". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.588419 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.588563 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09cfa50b-4138-4585-a53e-64dd3ab73335" (UID: "09cfa50b-4138-4585-a53e-64dd3ab73335"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.588938 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.614589 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "6ee8fbd3-1f81-4666-96da-5afc70819f1a" (UID: "6ee8fbd3-1f81-4666-96da-5afc70819f1a"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.614744 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities" (OuterVolumeSpecName: "utilities") pod "31fa8943-81cc-4750-a0b7-0fa9ab5af883" (UID: "31fa8943-81cc-4750-a0b7-0fa9ab5af883"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.589090 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.589355 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.588509 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.589598 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config" (OuterVolumeSpecName: "config") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.589617 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.589671 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6" (OuterVolumeSpecName: "kube-api-access-ftwb6") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "kube-api-access-ftwb6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.589683 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9" (OuterVolumeSpecName: "kube-api-access-ddlk9") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "kube-api-access-ddlk9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.581365 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities" (OuterVolumeSpecName: "utilities") pod "149b3c48-e17c-4a66-a835-d86dabf6ff13" (UID: "149b3c48-e17c-4a66-a835-d86dabf6ff13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.590000 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.590370 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d7e8f42f-dc0e-424b-bb56-5ec849834888" (UID: "d7e8f42f-dc0e-424b-bb56-5ec849834888"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.590451 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.590580 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.591674 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.591704 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m" (OuterVolumeSpecName: "kube-api-access-4hb7m") pod "94a6e063-3d1a-4d44-875d-185291448c31" (UID: "94a6e063-3d1a-4d44-875d-185291448c31"). InnerVolumeSpecName "kube-api-access-4hb7m". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592064 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images" (OuterVolumeSpecName: "images") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592092 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" (UID: "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.592798 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config" (OuterVolumeSpecName: "config") pod "7599e0b6-bddf-4def-b7f2-0b32206e8651" (UID: "7599e0b6-bddf-4def-b7f2-0b32206e8651"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595499 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh" (OuterVolumeSpecName: "kube-api-access-m5lgh") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "kube-api-access-m5lgh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.595803 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.596122 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.596292 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config" (OuterVolumeSpecName: "config") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.597596 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01080b46-74f1-4191-8755-5152a57b3b25" (UID: "01080b46-74f1-4191-8755-5152a57b3b25"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.615421 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.597640 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.597981 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.598025 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.598278 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.615624 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config" (OuterVolumeSpecName: "config") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.598771 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config" (OuterVolumeSpecName: "config") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.599530 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "42a11a02-47e1-488f-b270-2679d3298b0e" (UID: "42a11a02-47e1-488f-b270-2679d3298b0e"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.599789 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume" (OuterVolumeSpecName: "config-volume") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.600076 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.600109 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv" (OuterVolumeSpecName: "kube-api-access-6rmnv") pod "b605f283-6f2e-42da-a838-54421690f7d0" (UID: "b605f283-6f2e-42da-a838-54421690f7d0"). InnerVolumeSpecName "kube-api-access-6rmnv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.600139 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr" (OuterVolumeSpecName: "kube-api-access-z5rsr") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "kube-api-access-z5rsr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.600315 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.601506 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "81e39f7b-62e4-4fc9-992a-6535ce127a02" (UID: "81e39f7b-62e4-4fc9-992a-6535ce127a02"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.603042 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config" (OuterVolumeSpecName: "config") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.615727 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "b4750666-1362-4001-abd0-6f89964cc621" (UID: "b4750666-1362-4001-abd0-6f89964cc621"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.603059 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.603130 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities" (OuterVolumeSpecName: "utilities") pod "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" (UID: "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.603152 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.603672 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9" (OuterVolumeSpecName: "kube-api-access-99zj9") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "kube-api-access-99zj9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.603848 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.603605 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.604914 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.605221 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.605335 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn" (OuterVolumeSpecName: "kube-api-access-xnxbn") pod "ce090a97-9ab6-4c40-a719-64ff2acd9778" (UID: "ce090a97-9ab6-4c40-a719-64ff2acd9778"). InnerVolumeSpecName "kube-api-access-xnxbn". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.605708 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw" (OuterVolumeSpecName: "kube-api-access-5lcfw") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "kube-api-access-5lcfw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.605745 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.605945 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.606001 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.606979 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk" (OuterVolumeSpecName: "kube-api-access-qqbfk") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "kube-api-access-qqbfk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.606987 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.607142 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c" (OuterVolumeSpecName: "kube-api-access-8nb9c") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "kube-api-access-8nb9c". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.608034 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit" (OuterVolumeSpecName: "audit") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.609731 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr" (OuterVolumeSpecName: "kube-api-access-wj4qr") pod "149b3c48-e17c-4a66-a835-d86dabf6ff13" (UID: "149b3c48-e17c-4a66-a835-d86dabf6ff13"). InnerVolumeSpecName "kube-api-access-wj4qr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.611212 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert" (OuterVolumeSpecName: "cert") pod "a52afe44-fb37-46ed-a1f8-bf39727a3cbe" (UID: "a52afe44-fb37-46ed-a1f8-bf39727a3cbe"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.611365 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "e1d2a42d-af1d-4054-9618-ab545e0ed8b7" (UID: "e1d2a42d-af1d-4054-9618-ab545e0ed8b7"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.611808 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.612131 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.612517 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.612949 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc85e424-18b2-4924-920b-bd291a8c4b01" (UID: "cc85e424-18b2-4924-920b-bd291a8c4b01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.615180 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf" (OuterVolumeSpecName: "kube-api-access-nmmzf") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "kube-api-access-nmmzf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.616420 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t" (OuterVolumeSpecName: "kube-api-access-zth6t") pod "6077b63e-53a2-4f96-9d56-1ce0324e4913" (UID: "6077b63e-53a2-4f96-9d56-1ce0324e4913"). InnerVolumeSpecName "kube-api-access-zth6t". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.616581 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-proxy-tls\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.616623 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq" (OuterVolumeSpecName: "kube-api-access-d4tqq") pod "6ee8fbd3-1f81-4666-96da-5afc70819f1a" (UID: "6ee8fbd3-1f81-4666-96da-5afc70819f1a"). InnerVolumeSpecName "kube-api-access-d4tqq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.615177 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6" (OuterVolumeSpecName: "kube-api-access-pllx6") pod "81e39f7b-62e4-4fc9-992a-6535ce127a02" (UID: "81e39f7b-62e4-4fc9-992a-6535ce127a02"). InnerVolumeSpecName "kube-api-access-pllx6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.616712 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7" (OuterVolumeSpecName: "kube-api-access-tknt7") pod "584e1f4a-8205-47d7-8efb-3afc6017c4c9" (UID: "584e1f4a-8205-47d7-8efb-3afc6017c4c9"). InnerVolumeSpecName "kube-api-access-tknt7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.615764 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "f7e2c886-118e-43bb-bef1-c78134de392b" (UID: "f7e2c886-118e-43bb-bef1-c78134de392b"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.615950 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-rootfs\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.616852 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config" (OuterVolumeSpecName: "config") pod "01080b46-74f1-4191-8755-5152a57b3b25" (UID: "01080b46-74f1-4191-8755-5152a57b3b25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.617985 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl" (OuterVolumeSpecName: "kube-api-access-26xrl") pod "a208c9c2-333b-4b4a-be0d-bc32ec38a821" (UID: "a208c9c2-333b-4b4a-be0d-bc32ec38a821"). InnerVolumeSpecName "kube-api-access-26xrl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.618138 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp" (OuterVolumeSpecName: "tmp") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.618482 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities" (OuterVolumeSpecName: "utilities") pod "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" (UID: "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.618771 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp" (OuterVolumeSpecName: "tmp") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.621162 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-72hnj" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.622475 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap" (OuterVolumeSpecName: "whereabouts-flatfile-configmap") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "whereabouts-flatfile-configmap". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.622513 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.622521 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.622585 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.622684 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623154 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623349 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623491 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623285 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7" (OuterVolumeSpecName: "kube-api-access-hm9x7") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "kube-api-access-hm9x7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623326 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623317 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623580 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623604 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623637 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.623952 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.624592 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.625350 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz" (OuterVolumeSpecName: "kube-api-access-7jjkz") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "kube-api-access-7jjkz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.618749 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc" (OuterVolumeSpecName: "kube-api-access-zg8nc") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "kube-api-access-zg8nc". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.627226 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp" (OuterVolumeSpecName: "tmp") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.627584 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf" (OuterVolumeSpecName: "kube-api-access-6dmhf") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "kube-api-access-6dmhf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.627616 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "16bdd140-dce1-464c-ab47-dd5798d1d256" (UID: "16bdd140-dce1-464c-ab47-dd5798d1d256"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.627768 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.627791 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk" (OuterVolumeSpecName: "kube-api-access-w94wk") pod "01080b46-74f1-4191-8755-5152a57b3b25" (UID: "01080b46-74f1-4191-8755-5152a57b3b25"). InnerVolumeSpecName "kube-api-access-w94wk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.627946 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.628104 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.628103 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.628287 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h" (OuterVolumeSpecName: "kube-api-access-94l9h") pod "16bdd140-dce1-464c-ab47-dd5798d1d256" (UID: "16bdd140-dce1-464c-ab47-dd5798d1d256"). InnerVolumeSpecName "kube-api-access-94l9h". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.629598 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jq85\" (UniqueName: \"kubernetes.io/projected/6cb44ea3-ded3-4e25-a367-c6b9f0ab2153-kube-api-access-6jq85\") pod \"machine-config-daemon-w9wqq\" (UID: \"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\") " pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.630832 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.632045 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd" (OuterVolumeSpecName: "kube-api-access-mjwtd") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "kube-api-access-mjwtd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.632619 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.632790 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx" (OuterVolumeSpecName: "kube-api-access-l9stx") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "kube-api-access-l9stx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.633010 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d7e8f42f-dc0e-424b-bb56-5ec849834888" (UID: "d7e8f42f-dc0e-424b-bb56-5ec849834888"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.633892 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities" (OuterVolumeSpecName: "utilities") pod "94a6e063-3d1a-4d44-875d-185291448c31" (UID: "94a6e063-3d1a-4d44-875d-185291448c31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.634364 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca" (OuterVolumeSpecName: "service-ca") pod "d7e8f42f-dc0e-424b-bb56-5ec849834888" (UID: "d7e8f42f-dc0e-424b-bb56-5ec849834888"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.635643 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4sk9p" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.635647 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config" (OuterVolumeSpecName: "config") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.636645 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-44hwn\" (UniqueName: \"kubernetes.io/projected/e685d70e-7b85-464d-b251-f682148f431f-kube-api-access-44hwn\") pod \"multus-additional-cni-plugins-gzrb6\" (UID: \"e685d70e-7b85-464d-b251-f682148f431f\") " pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.638752 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg" (OuterVolumeSpecName: "kube-api-access-wbmqg") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "kube-api-access-wbmqg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.638769 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config" (OuterVolumeSpecName: "config") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.639050 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "81e39f7b-62e4-4fc9-992a-6535ce127a02" (UID: "81e39f7b-62e4-4fc9-992a-6535ce127a02"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.639334 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.640022 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config" (OuterVolumeSpecName: "config") pod "09cfa50b-4138-4585-a53e-64dd3ab73335" (UID: "09cfa50b-4138-4585-a53e-64dd3ab73335"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.640643 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w" (OuterVolumeSpecName: "kube-api-access-rzt4w") pod "a52afe44-fb37-46ed-a1f8-bf39727a3cbe" (UID: "a52afe44-fb37-46ed-a1f8-bf39727a3cbe"). InnerVolumeSpecName "kube-api-access-rzt4w". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.640899 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.641511 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities" (OuterVolumeSpecName: "utilities") pod "b605f283-6f2e-42da-a838-54421690f7d0" (UID: "b605f283-6f2e-42da-a838-54421690f7d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.641616 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl" (OuterVolumeSpecName: "kube-api-access-twvbl") pod "b4750666-1362-4001-abd0-6f89964cc621" (UID: "b4750666-1362-4001-abd0-6f89964cc621"). InnerVolumeSpecName "kube-api-access-twvbl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.641869 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem" (OuterVolumeSpecName: "ca-trust-extracted-pem") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "ca-trust-extracted-pem". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.641979 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.642125 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.643084 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs" (OuterVolumeSpecName: "kube-api-access-l87hs") pod "5ebfebf6-3ecd-458e-943f-bb25b52e2718" (UID: "5ebfebf6-3ecd-458e-943f-bb25b52e2718"). InnerVolumeSpecName "kube-api-access-l87hs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.643484 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-x6fv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a700740-bb4e-41e9-a9b3-516fc4546c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7k8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x6fv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.643930 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca" (OuterVolumeSpecName: "client-ca") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.644194 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "b4750666-1362-4001-abd0-6f89964cc621" (UID: "b4750666-1362-4001-abd0-6f89964cc621"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.644211 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config" (OuterVolumeSpecName: "config") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.647880 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "16bdd140-dce1-464c-ab47-dd5798d1d256" (UID: "16bdd140-dce1-464c-ab47-dd5798d1d256"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.648046 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images" (OuterVolumeSpecName: "images") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.648305 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.648322 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.648450 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.651383 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "149b3c48-e17c-4a66-a835-d86dabf6ff13" (UID: "149b3c48-e17c-4a66-a835-d86dabf6ff13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.651587 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.658212 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a068441-be86-4149-89e1-e1bb6bb0569b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcpq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcpq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-f4xkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.665234 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz" (OuterVolumeSpecName: "kube-api-access-ws8zz") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "kube-api-access-ws8zz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.665370 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l" (OuterVolumeSpecName: "kube-api-access-sbc2l") pod "593a3561-7760-45c5-8f91-5aaef7475d0f" (UID: "593a3561-7760-45c5-8f91-5aaef7475d0f"). InnerVolumeSpecName "kube-api-access-sbc2l". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.665383 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.665440 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.666921 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.666993 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.667017 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" (UID: "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: W0321 00:12:54.670571 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee6c0ddc_1c70_45c2_b50e_3cdce2f83288.slice/crio-916c968b02e334a935d502ad04119e1d611ab3dfa2d51d7a7e47163721517311 WatchSource:0}: Error finding container 916c968b02e334a935d502ad04119e1d611ab3dfa2d51d7a7e47163721517311: Status 404 returned error can't find the container with id 916c968b02e334a935d502ad04119e1d611ab3dfa2d51d7a7e47163721517311 Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.676616 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e685d70e-7b85-464d-b251-f682148f431f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gzrb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.686570 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.689733 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.693286 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94a6e063-3d1a-4d44-875d-185291448c31" (UID: "94a6e063-3d1a-4d44-875d-185291448c31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696313 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31fa8943-81cc-4750-a0b7-0fa9ab5af883" (UID: "31fa8943-81cc-4750-a0b7-0fa9ab5af883"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696873 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696902 5117 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696917 5117 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696935 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696948 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696961 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zg8nc\" (UniqueName: \"kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696974 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-sbc2l\" (UniqueName: \"kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696986 5117 reconciler_common.go:299] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.696999 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ddlk9\" (UniqueName: \"kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697013 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697026 5117 reconciler_common.go:299] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697040 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697054 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697066 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697079 5117 reconciler_common.go:299] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697092 5117 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697106 5117 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697120 5117 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697132 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697146 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6dmhf\" (UniqueName: \"kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697159 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697173 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-99zj9\" (UniqueName: \"kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697193 5117 reconciler_common.go:299] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697209 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xnxbn\" (UniqueName: \"kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697221 5117 reconciler_common.go:299] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697234 5117 reconciler_common.go:299] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697265 5117 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697279 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7jjkz\" (UniqueName: \"kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697290 5117 reconciler_common.go:299] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697301 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697313 5117 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697328 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pllx6\" (UniqueName: \"kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697342 5117 reconciler_common.go:299] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697355 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697368 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697381 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697394 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697407 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d4tqq\" (UniqueName: \"kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697419 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-94l9h\" (UniqueName: \"kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697431 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697445 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697458 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697471 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rzt4w\" (UniqueName: \"kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697484 5117 reconciler_common.go:299] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697497 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nmmzf\" (UniqueName: \"kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697515 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zth6t\" (UniqueName: \"kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697528 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8nb9c\" (UniqueName: \"kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697543 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wbmqg\" (UniqueName: \"kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697556 5117 reconciler_common.go:299] "Volume detached for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697571 5117 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697586 5117 reconciler_common.go:299] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697601 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tknt7\" (UniqueName: \"kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697615 5117 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697628 5117 reconciler_common.go:299] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697642 5117 reconciler_common.go:299] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697658 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697673 5117 reconciler_common.go:299] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697688 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l87hs\" (UniqueName: \"kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697702 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ws8zz\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697717 5117 reconciler_common.go:299] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697729 5117 reconciler_common.go:299] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697743 5117 reconciler_common.go:299] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697756 5117 reconciler_common.go:299] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697767 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wj4qr\" (UniqueName: \"kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697781 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-5lcfw\" (UniqueName: \"kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697794 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697808 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697820 5117 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697834 5117 reconciler_common.go:299] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697847 5117 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697859 5117 reconciler_common.go:299] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697870 5117 reconciler_common.go:299] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697884 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697899 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697913 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mjwtd\" (UniqueName: \"kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697928 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-w94wk\" (UniqueName: \"kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697940 5117 reconciler_common.go:299] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697956 5117 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697968 5117 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697981 5117 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.697995 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698007 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698020 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698032 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698049 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698062 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698077 5117 reconciler_common.go:299] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698091 5117 reconciler_common.go:299] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698103 5117 reconciler_common.go:299] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698116 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l9stx\" (UniqueName: \"kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698130 5117 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698141 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698154 5117 reconciler_common.go:299] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698165 5117 reconciler_common.go:299] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698178 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xfp5s\" (UniqueName: \"kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698192 5117 reconciler_common.go:299] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698205 5117 reconciler_common.go:299] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698217 5117 reconciler_common.go:299] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698231 5117 reconciler_common.go:299] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698244 5117 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698275 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698287 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698302 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698317 5117 reconciler_common.go:299] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698333 5117 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698347 5117 reconciler_common.go:299] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698363 5117 reconciler_common.go:299] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698377 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698391 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698404 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698419 5117 reconciler_common.go:299] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698433 5117 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698445 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698458 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698471 5117 reconciler_common.go:299] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698483 5117 reconciler_common.go:299] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698497 5117 reconciler_common.go:299] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698510 5117 reconciler_common.go:299] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698522 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-z5rsr\" (UniqueName: \"kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698536 5117 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698550 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ftwb6\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698563 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4hb7m\" (UniqueName: \"kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698578 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-m5lgh\" (UniqueName: \"kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698591 5117 reconciler_common.go:299] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698605 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698617 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698629 5117 reconciler_common.go:299] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698641 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hm9x7\" (UniqueName: \"kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698654 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-26xrl\" (UniqueName: \"kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698667 5117 reconciler_common.go:299] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698678 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6rmnv\" (UniqueName: \"kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698690 5117 reconciler_common.go:299] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698702 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698714 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698726 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qqbfk\" (UniqueName: \"kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698740 5117 reconciler_common.go:299] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698751 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698763 5117 reconciler_common.go:299] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698774 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698791 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698803 5117 reconciler_common.go:299] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698814 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698825 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698838 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698851 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-twvbl\" (UniqueName: \"kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698866 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698878 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698889 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698901 5117 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698914 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698926 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698939 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698953 5117 reconciler_common.go:299] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698966 5117 reconciler_common.go:299] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.698977 5117 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.707751 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" (UID: "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.710772 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9kggs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: W0321 00:12:54.715892 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode685d70e_7b85_464d_b251_f682148f431f.slice/crio-67eb4695d4fcea11a98239436dd4281d47925773f438bb7214f7ccc002668824 WatchSource:0}: Error finding container 67eb4695d4fcea11a98239436dd4281d47925773f438bb7214f7ccc002668824: Status 404 returned error can't find the container with id 67eb4695d4fcea11a98239436dd4281d47925773f438bb7214f7ccc002668824 Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.717166 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.726548 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.726589 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.726604 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.726624 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.726639 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: W0321 00:12:54.730600 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdb18fcd_7135_4474_ada3_a29788223262.slice/crio-0e6e9972da340b3c5fff318e95cafef52b9aed66f1e9e21bd05d1e3c232c6a24 WatchSource:0}: Error finding container 0e6e9972da340b3c5fff318e95cafef52b9aed66f1e9e21bd05d1e3c232c6a24: Status 404 returned error can't find the container with id 0e6e9972da340b3c5fff318e95cafef52b9aed66f1e9e21bd05d1e3c232c6a24 Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.799436 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.830852 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.830886 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.830896 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.830911 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.830924 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.900738 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.900785 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.900809 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.900829 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.900916 5117 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.900976 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:55.900962222 +0000 UTC m=+89.194249384 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901541 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901566 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901579 5117 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901608 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:55.901600142 +0000 UTC m=+89.194887314 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901650 5117 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901674 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:55.901666944 +0000 UTC m=+89.194954116 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901714 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901724 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901731 5117 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:54 crc kubenswrapper[5117]: E0321 00:12:54.901750 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:55.901744136 +0000 UTC m=+89.195031308 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.933991 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.934051 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.934064 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.934083 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.934095 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:54Z","lastTransitionTime":"2026-03-21T00:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.940625 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" event={"ID":"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc","Type":"ContainerStarted","Data":"fd8f9283b9b2a2352c740526544cd3f64eb105d8698e87135c3d1eb768f1c96a"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.943062 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" event={"ID":"fc4541ce-7789-4670-bc75-5c2868e52ce0","Type":"ContainerStarted","Data":"5ee679876dff16f3c1225dfcdde8b7dc769373444df26682e67a51103548316b"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.943090 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" event={"ID":"fc4541ce-7789-4670-bc75-5c2868e52ce0","Type":"ContainerStarted","Data":"f267fac5a93802824762b8a2a59b78bb435633b8df58049bc07de79086f53bcb"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.945656 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerID="1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34" exitCode=0 Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.945705 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.945722 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerStarted","Data":"916c968b02e334a935d502ad04119e1d611ab3dfa2d51d7a7e47163721517311"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.956879 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-x6fv6" event={"ID":"3a700740-bb4e-41e9-a9b3-516fc4546c5a","Type":"ContainerStarted","Data":"a81cd166149f657056763af163580ebdcb74736737b3fd815cabb12777753b3a"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.958292 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e2c803-23f7-446f-a7f3-90de8dc4355c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://af1c6802eebfabdf424eae7ad9eecfcaaffcd44a0d2cfad7300f2ff7a532cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://3a5498baab91a2da6643fcdc8b35141226196ffe302f6c2f38a266e518799587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://cb3b6dc02877e4af1eb1e9c2de8b9041ba203d71b6e52185287a6cd4c3ba28b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c6f7650b6653e951c6fe6e6883d205ee80691d8683967540c1766ef0d1df7f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6f7650b6653e951c6fe6e6883d205ee80691d8683967540c1766ef0d1df7f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.961557 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4sk9p" event={"ID":"bdb18fcd-7135-4474-ada3-a29788223262","Type":"ContainerStarted","Data":"0e6e9972da340b3c5fff318e95cafef52b9aed66f1e9e21bd05d1e3c232c6a24"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.969269 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-72hnj" event={"ID":"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e","Type":"ContainerStarted","Data":"3f68e394d4ab5e40e8dc7c669538f180f3f69df9c5011076f34f1723b3472c4a"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.969365 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-72hnj" event={"ID":"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e","Type":"ContainerStarted","Data":"67ab1c7359099449e4085e545f366986af8863986a34ac05d743e46bad8a2a5d"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.971741 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" event={"ID":"34177974-8d82-49d2-a763-391d0df3bbd8","Type":"ContainerStarted","Data":"7812f5ccb08fd13b40438fab151c717a49aa8aeaf5eaa68a8e45a942b039fc9c"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.971787 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" event={"ID":"34177974-8d82-49d2-a763-391d0df3bbd8","Type":"ContainerStarted","Data":"969e55eb73049032c3b8f49bcecae7efa8dc9fa359e490a19d6cafebcea05fef"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.974566 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerStarted","Data":"5d05bbaf9f43d60a457e320460b8471e1a6260b3ff7dfd5564f0a0339b2cd61d"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.975863 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerStarted","Data":"67eb4695d4fcea11a98239436dd4281d47925773f438bb7214f7ccc002668824"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.976344 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797d5d7c-c536-4487-99be-f1f0219b63c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T00:12:42Z\\\",\\\"message\\\":\\\"vvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nW0321 00:12:42.499742 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 00:12:42.499925 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI0321 00:12:42.501331 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3355725928/tls.crt::/tmp/serving-cert-3355725928/tls.key\\\\\\\"\\\\nI0321 00:12:42.934594 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 00:12:42.943210 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 00:12:42.943284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 00:12:42.943339 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 00:12:42.943351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 00:12:42.951353 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 00:12:42.951415 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 00:12:42.951428 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 00:12:42.951440 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0321 00:12:42.951448 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 00:12:42.951457 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 00:12:42.951465 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0321 00:12:42.951439 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0321 00:12:42.954854 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T00:12:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.977332 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" event={"ID":"1a068441-be86-4149-89e1-e1bb6bb0569b","Type":"ContainerStarted","Data":"114386f51b93bcf86a51cb7dabf09cd927daa2b4b93cde30c17bade4a5bfbe87"} Mar 21 00:12:54 crc kubenswrapper[5117]: I0321 00:12:54.989066 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.001458 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.001842 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.002410 5117 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.002545 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs podName:482a0770-6dc5-4682-bf1e-e3c085129f33 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:56.002518714 +0000 UTC m=+89.295806086 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs") pod "network-metrics-daemon-qp2t9" (UID: "482a0770-6dc5-4682-bf1e-e3c085129f33") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.012922 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-x6fv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a700740-bb4e-41e9-a9b3-516fc4546c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7k8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x6fv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.027372 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a068441-be86-4149-89e1-e1bb6bb0569b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcpq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcpq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-f4xkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.035975 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.036035 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.036047 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.036063 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.036073 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.042476 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e685d70e-7b85-464d-b251-f682148f431f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gzrb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.053822 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.079391 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"resources\\\":{},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:12:54Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9kggs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.094322 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d4e366f-7f9a-45ea-9ccd-505c06a20f24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://94684ef43a4caf326c026a0b8b43fe90d3f86c9c60562f24fa305fcfc6f70788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://ad24a3d986e06d5c3f1905aa7e44d8fcc888cb7acc101ca3cbb39adac8684500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://d9eee0d1d08899481bf9b5b0fe5429a679042a29b12e717e65d1a11a9ab2d79e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ad515cb5c17c39877403cc2a002a95cf96b2d32f46c7cb4b753d390ad268c3dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.103738 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.104926 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:12:56.104893491 +0000 UTC m=+89.398180663 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.107077 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7b9c1f1-3a7c-4cad-9c70-09dc9f7e75b5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c2514a311c8df4b1faca8a81fa4c3becfae5f7e05e32fe441cc7d61e4b3fbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://000182f0824f8c8f7487c695f547ec126efb5f50836571807732678886484fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://000182f0824f8c8f7487c695f547ec126efb5f50836571807732678886484fe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.126816 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eae805bb-c929-4ab2-af7b-3e582255fe6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"},\\\"containerID\\\":\\\"cri-o://f6e2280de7b69beabb5b164fe42e7440579b41b13e07f84fcd7a6ad773218de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://825d8b369cbcdc65a64926c868202d0b3b9253162f40994fa0f447a5beeede1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ed17e61de6c00e7e405bc7beaa0ce1f7e26b6a4ec68e83ebee1f4c7ec0d3f27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b743995a17c55f75fba00e6815f49b6fc598d27f4c4c81452bad9ca5806d9f80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://b44d38825bd92498e91533afc0db62985336c2b9293ccdc5221262dcacd0acde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:31Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://1f7effd8d591e2ac11bcdb21df51a7b48ab90a589f431a721326bea14181aa6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7effd8d591e2ac11bcdb21df51a7b48ab90a589f431a721326bea14181aa6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd-auto-backup\\\",\\\"name\\\":\\\"etcd-auto-backup-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://c2f608613c2ee33e9d147e0481d3f47512c4db2c0e5fe416c84ad073975f3d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f608613c2ee33e9d147e0481d3f47512c4db2c0e5fe416c84ad073975f3d5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://d6e7a9fe3510c2f332f71cb736921b49b751c42424d2cbb5f6e9cb6c3726555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e7a9fe3510c2f332f71cb736921b49b751c42424d2cbb5f6e9cb6c3726555e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.137421 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.138226 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.138281 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.138292 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.138310 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.138321 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.148663 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.158867 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.176056 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-72hnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ktw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-72hnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.187877 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qp2t9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482a0770-6dc5-4682-bf1e-e3c085129f33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42dc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42dc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qp2t9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.199109 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4sk9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdb18fcd-7135-4474-ada3-a29788223262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nlcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4sk9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.210845 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jq85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jq85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9wqq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.225515 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jq85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jq85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9wqq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.239480 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e2c803-23f7-446f-a7f3-90de8dc4355c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://af1c6802eebfabdf424eae7ad9eecfcaaffcd44a0d2cfad7300f2ff7a532cadf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://3a5498baab91a2da6643fcdc8b35141226196ffe302f6c2f38a266e518799587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://cb3b6dc02877e4af1eb1e9c2de8b9041ba203d71b6e52185287a6cd4c3ba28b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c6f7650b6653e951c6fe6e6883d205ee80691d8683967540c1766ef0d1df7f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6f7650b6653e951c6fe6e6883d205ee80691d8683967540c1766ef0d1df7f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.250837 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.250901 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.250918 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.250939 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.250959 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.273771 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"797d5d7c-c536-4487-99be-f1f0219b63c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-21T00:12:42Z\\\",\\\"message\\\":\\\"vvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nW0321 00:12:42.499742 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0321 00:12:42.499925 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI0321 00:12:42.501331 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3355725928/tls.crt::/tmp/serving-cert-3355725928/tls.key\\\\\\\"\\\\nI0321 00:12:42.934594 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0321 00:12:42.943210 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0321 00:12:42.943284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0321 00:12:42.943339 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0321 00:12:42.943351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0321 00:12:42.951353 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0321 00:12:42.951415 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 00:12:42.951428 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0321 00:12:42.951440 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0321 00:12:42.951448 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0321 00:12:42.951457 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0321 00:12:42.951465 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0321 00:12:42.951439 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0321 00:12:42.954854 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-21T00:12:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.312961 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://7812f5ccb08fd13b40438fab151c717a49aa8aeaf5eaa68a8e45a942b039fc9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:12:54Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.348636 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.353530 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.353574 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.353584 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.353599 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.353608 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.387012 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-x6fv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a700740-bb4e-41e9-a9b3-516fc4546c5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j7k8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-x6fv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.432024 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a068441-be86-4149-89e1-e1bb6bb0569b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcpq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcpq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-f4xkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.457055 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.457131 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.457142 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.457159 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.457174 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.469934 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e685d70e-7b85-464d-b251-f682148f431f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44hwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gzrb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.507630 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.555901 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"resources\\\":{},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:12:54Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmjsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9kggs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.559989 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.560062 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.560083 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.560103 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.560115 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.579044 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.579166 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.579320 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.579349 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qp2t9" podUID="482a0770-6dc5-4682-bf1e-e3c085129f33" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.579428 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.579537 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.579576 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.579682 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.587932 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01080b46-74f1-4191-8755-5152a57b3b25" path="/var/lib/kubelet/pods/01080b46-74f1-4191-8755-5152a57b3b25/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.589059 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09cfa50b-4138-4585-a53e-64dd3ab73335" path="/var/lib/kubelet/pods/09cfa50b-4138-4585-a53e-64dd3ab73335/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.592313 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dd0fbac-8c0d-4228-8faa-abbeedabf7db" path="/var/lib/kubelet/pods/0dd0fbac-8c0d-4228-8faa-abbeedabf7db/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.596026 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0effdbcf-dd7d-404d-9d48-77536d665a5d" path="/var/lib/kubelet/pods/0effdbcf-dd7d-404d-9d48-77536d665a5d/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.600085 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="149b3c48-e17c-4a66-a835-d86dabf6ff13" path="/var/lib/kubelet/pods/149b3c48-e17c-4a66-a835-d86dabf6ff13/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.600971 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d4e366f-7f9a-45ea-9ccd-505c06a20f24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://94684ef43a4caf326c026a0b8b43fe90d3f86c9c60562f24fa305fcfc6f70788\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://ad24a3d986e06d5c3f1905aa7e44d8fcc888cb7acc101ca3cbb39adac8684500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://d9eee0d1d08899481bf9b5b0fe5429a679042a29b12e717e65d1a11a9ab2d79e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ad515cb5c17c39877403cc2a002a95cf96b2d32f46c7cb4b753d390ad268c3dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.605079 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16bdd140-dce1-464c-ab47-dd5798d1d256" path="/var/lib/kubelet/pods/16bdd140-dce1-464c-ab47-dd5798d1d256/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.606172 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18f80adb-c1c3-49ba-8ee4-932c851d3897" path="/var/lib/kubelet/pods/18f80adb-c1c3-49ba-8ee4-932c851d3897/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.607911 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" path="/var/lib/kubelet/pods/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.608677 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2325ffef-9d5b-447f-b00e-3efc429acefe" path="/var/lib/kubelet/pods/2325ffef-9d5b-447f-b00e-3efc429acefe/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.611599 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="301e1965-1754-483d-b6cc-bfae7038bbca" path="/var/lib/kubelet/pods/301e1965-1754-483d-b6cc-bfae7038bbca/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.614427 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31fa8943-81cc-4750-a0b7-0fa9ab5af883" path="/var/lib/kubelet/pods/31fa8943-81cc-4750-a0b7-0fa9ab5af883/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.616554 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a11a02-47e1-488f-b270-2679d3298b0e" path="/var/lib/kubelet/pods/42a11a02-47e1-488f-b270-2679d3298b0e/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.617760 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="567683bd-0efc-4f21-b076-e28559628404" path="/var/lib/kubelet/pods/567683bd-0efc-4f21-b076-e28559628404/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.620652 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="584e1f4a-8205-47d7-8efb-3afc6017c4c9" path="/var/lib/kubelet/pods/584e1f4a-8205-47d7-8efb-3afc6017c4c9/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.621208 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="593a3561-7760-45c5-8f91-5aaef7475d0f" path="/var/lib/kubelet/pods/593a3561-7760-45c5-8f91-5aaef7475d0f/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.622624 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ebfebf6-3ecd-458e-943f-bb25b52e2718" path="/var/lib/kubelet/pods/5ebfebf6-3ecd-458e-943f-bb25b52e2718/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.623373 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6077b63e-53a2-4f96-9d56-1ce0324e4913" path="/var/lib/kubelet/pods/6077b63e-53a2-4f96-9d56-1ce0324e4913/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.625064 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" path="/var/lib/kubelet/pods/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.626841 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6edfcf45-925b-4eff-b940-95b6fc0b85d4" path="/var/lib/kubelet/pods/6edfcf45-925b-4eff-b940-95b6fc0b85d4/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.628393 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7b9c1f1-3a7c-4cad-9c70-09dc9f7e75b5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c2514a311c8df4b1faca8a81fa4c3becfae5f7e05e32fe441cc7d61e4b3fbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://000182f0824f8c8f7487c695f547ec126efb5f50836571807732678886484fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://000182f0824f8c8f7487c695f547ec126efb5f50836571807732678886484fe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.628555 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ee8fbd3-1f81-4666-96da-5afc70819f1a" path="/var/lib/kubelet/pods/6ee8fbd3-1f81-4666-96da-5afc70819f1a/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.630791 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" path="/var/lib/kubelet/pods/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.633673 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="736c54fe-349c-4bb9-870a-d1c1d1c03831" path="/var/lib/kubelet/pods/736c54fe-349c-4bb9-870a-d1c1d1c03831/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.635057 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7599e0b6-bddf-4def-b7f2-0b32206e8651" path="/var/lib/kubelet/pods/7599e0b6-bddf-4def-b7f2-0b32206e8651/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.636776 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7afa918d-be67-40a6-803c-d3b0ae99d815" path="/var/lib/kubelet/pods/7afa918d-be67-40a6-803c-d3b0ae99d815/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.638065 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df94c10-441d-4386-93a6-6730fb7bcde0" path="/var/lib/kubelet/pods/7df94c10-441d-4386-93a6-6730fb7bcde0/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.639236 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" path="/var/lib/kubelet/pods/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.642720 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e39f7b-62e4-4fc9-992a-6535ce127a02" path="/var/lib/kubelet/pods/81e39f7b-62e4-4fc9-992a-6535ce127a02/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.643781 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869851b9-7ffb-4af0-b166-1d8aa40a5f80" path="/var/lib/kubelet/pods/869851b9-7ffb-4af0-b166-1d8aa40a5f80/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.647714 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" path="/var/lib/kubelet/pods/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.648233 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92dfbade-90b6-4169-8c07-72cff7f2c82b" path="/var/lib/kubelet/pods/92dfbade-90b6-4169-8c07-72cff7f2c82b/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.652122 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a6e063-3d1a-4d44-875d-185291448c31" path="/var/lib/kubelet/pods/94a6e063-3d1a-4d44-875d-185291448c31/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.656805 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f71a554-e414-4bc3-96d2-674060397afe" path="/var/lib/kubelet/pods/9f71a554-e414-4bc3-96d2-674060397afe/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.666449 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.666501 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.666517 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.666537 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.666552 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.666584 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a208c9c2-333b-4b4a-be0d-bc32ec38a821" path="/var/lib/kubelet/pods/a208c9c2-333b-4b4a-be0d-bc32ec38a821/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.670036 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a52afe44-fb37-46ed-a1f8-bf39727a3cbe" path="/var/lib/kubelet/pods/a52afe44-fb37-46ed-a1f8-bf39727a3cbe/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.671113 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a555ff2e-0be6-46d5-897d-863bb92ae2b3" path="/var/lib/kubelet/pods/a555ff2e-0be6-46d5-897d-863bb92ae2b3/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.671699 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a88189-c967-4640-879e-27665747f20c" path="/var/lib/kubelet/pods/a7a88189-c967-4640-879e-27665747f20c/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.673550 5117 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="af33e427-6803-48c2-a76a-dd9deb7cbf9a" path="/var/lib/kubelet/pods/af33e427-6803-48c2-a76a-dd9deb7cbf9a/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.673692 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af33e427-6803-48c2-a76a-dd9deb7cbf9a" path="/var/lib/kubelet/pods/af33e427-6803-48c2-a76a-dd9deb7cbf9a/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.679622 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af41de71-79cf-4590-bbe9-9e8b848862cb" path="/var/lib/kubelet/pods/af41de71-79cf-4590-bbe9-9e8b848862cb/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.683464 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" path="/var/lib/kubelet/pods/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.685418 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4750666-1362-4001-abd0-6f89964cc621" path="/var/lib/kubelet/pods/b4750666-1362-4001-abd0-6f89964cc621/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.686883 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b605f283-6f2e-42da-a838-54421690f7d0" path="/var/lib/kubelet/pods/b605f283-6f2e-42da-a838-54421690f7d0/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.687407 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c491984c-7d4b-44aa-8c1e-d7974424fa47" path="/var/lib/kubelet/pods/c491984c-7d4b-44aa-8c1e-d7974424fa47/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.689236 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5f2bfad-70f6-4185-a3d9-81ce12720767" path="/var/lib/kubelet/pods/c5f2bfad-70f6-4185-a3d9-81ce12720767/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.690181 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc85e424-18b2-4924-920b-bd291a8c4b01" path="/var/lib/kubelet/pods/cc85e424-18b2-4924-920b-bd291a8c4b01/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.691503 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce090a97-9ab6-4c40-a719-64ff2acd9778" path="/var/lib/kubelet/pods/ce090a97-9ab6-4c40-a719-64ff2acd9778/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.693720 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eae805bb-c929-4ab2-af7b-3e582255fe6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"},\\\"containerID\\\":\\\"cri-o://f6e2280de7b69beabb5b164fe42e7440579b41b13e07f84fcd7a6ad773218de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://825d8b369cbcdc65a64926c868202d0b3b9253162f40994fa0f447a5beeede1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ed17e61de6c00e7e405bc7beaa0ce1f7e26b6a4ec68e83ebee1f4c7ec0d3f27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b743995a17c55f75fba00e6815f49b6fc598d27f4c4c81452bad9ca5806d9f80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:32Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://b44d38825bd92498e91533afc0db62985336c2b9293ccdc5221262dcacd0acde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:11:31Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://1f7effd8d591e2ac11bcdb21df51a7b48ab90a589f431a721326bea14181aa6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f7effd8d591e2ac11bcdb21df51a7b48ab90a589f431a721326bea14181aa6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:28Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd-auto-backup\\\",\\\"name\\\":\\\"etcd-auto-backup-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://c2f608613c2ee33e9d147e0481d3f47512c4db2c0e5fe416c84ad073975f3d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2f608613c2ee33e9d147e0481d3f47512c4db2c0e5fe416c84ad073975f3d5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:29Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://d6e7a9fe3510c2f332f71cb736921b49b751c42424d2cbb5f6e9cb6c3726555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e7a9fe3510c2f332f71cb736921b49b751c42424d2cbb5f6e9cb6c3726555e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-21T00:11:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-21T00:11:30Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:11:27Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.694168 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d19cb085-0c5b-4810-b654-ce7923221d90" path="/var/lib/kubelet/pods/d19cb085-0c5b-4810-b654-ce7923221d90/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.695960 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" path="/var/lib/kubelet/pods/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.706563 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d565531a-ff86-4608-9d19-767de01ac31b" path="/var/lib/kubelet/pods/d565531a-ff86-4608-9d19-767de01ac31b/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.707441 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e8f42f-dc0e-424b-bb56-5ec849834888" path="/var/lib/kubelet/pods/d7e8f42f-dc0e-424b-bb56-5ec849834888/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.718554 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.725122 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" path="/var/lib/kubelet/pods/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.737211 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e093be35-bb62-4843-b2e8-094545761610" path="/var/lib/kubelet/pods/e093be35-bb62-4843-b2e8-094545761610/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.738057 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" path="/var/lib/kubelet/pods/e1d2a42d-af1d-4054-9618-ab545e0ed8b7/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.738998 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f559dfa3-3917-43a2-97f6-61ddfda10e93" path="/var/lib/kubelet/pods/f559dfa3-3917-43a2-97f6-61ddfda10e93/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.750075 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65c0ac1-8bca-454d-a2e6-e35cb418beac" path="/var/lib/kubelet/pods/f65c0ac1-8bca-454d-a2e6-e35cb418beac/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.750834 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.751161 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" path="/var/lib/kubelet/pods/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.751989 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e2c886-118e-43bb-bef1-c78134de392b" path="/var/lib/kubelet/pods/f7e2c886-118e-43bb-bef1-c78134de392b/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.768284 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.768339 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.768352 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.768371 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.768386 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.773106 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" path="/var/lib/kubelet/pods/fc8db2c7-859d-47b3-a900-2bd0c0b2973b/volumes" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.790785 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.829830 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-72hnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"65Mi\\\"},\\\"containerID\\\":\\\"cri-o://3f68e394d4ab5e40e8dc7c669538f180f3f69df9c5011076f34f1723b3472c4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"65Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-21T00:12:54Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2ktw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-72hnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.868823 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qp2t9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482a0770-6dc5-4682-bf1e-e3c085129f33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42dc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-42dc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qp2t9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.871807 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.871896 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.871916 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.871942 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.872156 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.909031 5117 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4sk9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdb18fcd-7135-4474-ada3-a29788223262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-21T00:12:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2nlcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-21T00:12:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4sk9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.916810 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.917094 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.917123 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.917158 5117 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.917272 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:57.917209955 +0000 UTC m=+91.210497137 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.917335 5117 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.917528 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:57.917481533 +0000 UTC m=+91.210768745 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.917874 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.917950 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.918006 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.918130 5117 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.918199 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:57.918188485 +0000 UTC m=+91.211475677 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.918332 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.918348 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.918361 5117 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:55 crc kubenswrapper[5117]: E0321 00:12:55.918418 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:57.91838641 +0000 UTC m=+91.211673602 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.974503 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.974549 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.974560 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.974575 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.974588 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:55Z","lastTransitionTime":"2026-03-21T00:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.983756 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" event={"ID":"1a068441-be86-4149-89e1-e1bb6bb0569b","Type":"ContainerStarted","Data":"265ada799a3ab9199d0147cd05bd4d41a3acf3ba376cbef5fba0041b69d968cc"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.983810 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" event={"ID":"1a068441-be86-4149-89e1-e1bb6bb0569b","Type":"ContainerStarted","Data":"f64a733df56c6d6741a0fd293d4e4fa09f79855c173e321da1ba6b1fc5e80390"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.986846 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" event={"ID":"fc4541ce-7789-4670-bc75-5c2868e52ce0","Type":"ContainerStarted","Data":"a29c5aee193ddc78315021e228988e567b0b4f3217962fc9904d929c71c6b182"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.994925 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerStarted","Data":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.994961 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerStarted","Data":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.994971 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerStarted","Data":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.994980 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerStarted","Data":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.996329 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-x6fv6" event={"ID":"3a700740-bb4e-41e9-a9b3-516fc4546c5a","Type":"ContainerStarted","Data":"c3b9a88623579a4eba458c669fda76ec0292f9b24e153ff0f6a03c3bdea4b3d4"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.998169 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4sk9p" event={"ID":"bdb18fcd-7135-4474-ada3-a29788223262","Type":"ContainerStarted","Data":"f433767d1c4305ebd113458818e788568962c2c10b56700d7b659e59d5ffad7b"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.999576 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerStarted","Data":"30feb01bffab75ff5502bd18d09e128ff61d52b61951365e37a25119843f62b7"} Mar 21 00:12:55 crc kubenswrapper[5117]: I0321 00:12:55.999595 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerStarted","Data":"4c8552401b305d82d54f228c0905aa24fe71ef36fb91e0372632f7b632cbf5e9"} Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.002384 5117 generic.go:358] "Generic (PLEG): container finished" podID="e685d70e-7b85-464d-b251-f682148f431f" containerID="683ef3cd4261697ab799ebf2439db2d9787328fd10fdfa4ba96d5eab3c5ee301" exitCode=0 Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.002459 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerDied","Data":"683ef3cd4261697ab799ebf2439db2d9787328fd10fdfa4ba96d5eab3c5ee301"} Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.019297 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:56 crc kubenswrapper[5117]: E0321 00:12:56.020849 5117 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:12:56 crc kubenswrapper[5117]: E0321 00:12:56.020959 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs podName:482a0770-6dc5-4682-bf1e-e3c085129f33 nodeName:}" failed. No retries permitted until 2026-03-21 00:12:58.020935363 +0000 UTC m=+91.314222545 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs") pod "network-metrics-daemon-qp2t9" (UID: "482a0770-6dc5-4682-bf1e-e3c085129f33") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.023104 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.023082939 podStartE2EDuration="3.023082939s" podCreationTimestamp="2026-03-21 00:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:56.02081473 +0000 UTC m=+89.314101912" watchObservedRunningTime="2026-03-21 00:12:56.023082939 +0000 UTC m=+89.316370111" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.078584 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.079233 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.079270 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.079287 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.079300 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:56Z","lastTransitionTime":"2026-03-21T00:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.121510 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:56 crc kubenswrapper[5117]: E0321 00:12:56.123347 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:12:58.12331562 +0000 UTC m=+91.416602802 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.150088 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-72hnj" podStartSLOduration=68.150066278 podStartE2EDuration="1m8.150066278s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:56.113203542 +0000 UTC m=+89.406490734" watchObservedRunningTime="2026-03-21 00:12:56.150066278 +0000 UTC m=+89.443353450" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.160957 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.161010 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.161022 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.161042 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.161054 5117 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-21T00:12:56Z","lastTransitionTime":"2026-03-21T00:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.202657 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm"] Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.210703 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.242664 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"default-dockercfg-hqpm5\"" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.262895 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"cluster-version-operator-serving-cert\"" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.281960 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"openshift-service-ca.crt\"" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.302642 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"kube-root-ca.crt\"" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.323287 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d1770ea5-3738-4407-8de2-c4d3aea8385e-service-ca\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.323342 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d1770ea5-3738-4407-8de2-c4d3aea8385e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.323365 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d1770ea5-3738-4407-8de2-c4d3aea8385e-etc-ssl-certs\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.323540 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1770ea5-3738-4407-8de2-c4d3aea8385e-serving-cert\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.323619 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1770ea5-3738-4407-8de2-c4d3aea8385e-kube-api-access\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.399303 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=3.399278201 podStartE2EDuration="3.399278201s" podCreationTimestamp="2026-03-21 00:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:56.354379109 +0000 UTC m=+89.647666281" watchObservedRunningTime="2026-03-21 00:12:56.399278201 +0000 UTC m=+89.692565383" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.424495 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d1770ea5-3738-4407-8de2-c4d3aea8385e-service-ca\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.424563 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d1770ea5-3738-4407-8de2-c4d3aea8385e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.424592 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d1770ea5-3738-4407-8de2-c4d3aea8385e-etc-ssl-certs\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.424638 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1770ea5-3738-4407-8de2-c4d3aea8385e-serving-cert\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.424663 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1770ea5-3738-4407-8de2-c4d3aea8385e-kube-api-access\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.425514 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d1770ea5-3738-4407-8de2-c4d3aea8385e-service-ca\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.425567 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d1770ea5-3738-4407-8de2-c4d3aea8385e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.425594 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d1770ea5-3738-4407-8de2-c4d3aea8385e-etc-ssl-certs\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.445671 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1770ea5-3738-4407-8de2-c4d3aea8385e-serving-cert\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.456048 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1770ea5-3738-4407-8de2-c4d3aea8385e-kube-api-access\") pod \"cluster-version-operator-7c9b9cfd6-lmfdm\" (UID: \"d1770ea5-3738-4407-8de2-c4d3aea8385e\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.524357 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.542667 5117 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Mar 21 00:12:56 crc kubenswrapper[5117]: W0321 00:12:56.543912 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1770ea5_3738_4407_8de2_c4d3aea8385e.slice/crio-161a4594b80f84739878451add08dc3daec514a724c67dd83b14c69be7c7ac35 WatchSource:0}: Error finding container 161a4594b80f84739878451add08dc3daec514a724c67dd83b14c69be7c7ac35: Status 404 returned error can't find the container with id 161a4594b80f84739878451add08dc3daec514a724c67dd83b14c69be7c7ac35 Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.596592 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" podStartSLOduration=68.596555096 podStartE2EDuration="1m8.596555096s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:56.595470954 +0000 UTC m=+89.888758136" watchObservedRunningTime="2026-03-21 00:12:56.596555096 +0000 UTC m=+89.889842298" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.642263 5117 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.795082 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=3.795063 podStartE2EDuration="3.795063s" podCreationTimestamp="2026-03-21 00:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:56.794775961 +0000 UTC m=+90.088063143" watchObservedRunningTime="2026-03-21 00:12:56.795063 +0000 UTC m=+90.088350172" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.832344 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=3.832324269 podStartE2EDuration="3.832324269s" podCreationTimestamp="2026-03-21 00:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:56.831378459 +0000 UTC m=+90.124665631" watchObservedRunningTime="2026-03-21 00:12:56.832324269 +0000 UTC m=+90.125611441" Mar 21 00:12:56 crc kubenswrapper[5117]: I0321 00:12:56.873398 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-x6fv6" podStartSLOduration=68.873380472 podStartE2EDuration="1m8.873380472s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:56.872810506 +0000 UTC m=+90.166097668" watchObservedRunningTime="2026-03-21 00:12:56.873380472 +0000 UTC m=+90.166667644" Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.008632 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerStarted","Data":"9da9ecf69ebeb25146b645d2c60cec6840b1730edf09ecef607181a9dd25e8dc"} Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.009332 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" event={"ID":"d1770ea5-3738-4407-8de2-c4d3aea8385e","Type":"ContainerStarted","Data":"161a4594b80f84739878451add08dc3daec514a724c67dd83b14c69be7c7ac35"} Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.012639 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerStarted","Data":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.012669 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerStarted","Data":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.074242 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podStartSLOduration=69.074216148 podStartE2EDuration="1m9.074216148s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:57.073342841 +0000 UTC m=+90.366630013" watchObservedRunningTime="2026-03-21 00:12:57.074216148 +0000 UTC m=+90.367503320" Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.075188 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-4sk9p" podStartSLOduration=69.075180137 podStartE2EDuration="1m9.075180137s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:57.035436233 +0000 UTC m=+90.328723415" watchObservedRunningTime="2026-03-21 00:12:57.075180137 +0000 UTC m=+90.368467309" Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.579807 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.580329 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.580100 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.580418 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.580113 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.580484 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.580071 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.580541 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qp2t9" podUID="482a0770-6dc5-4682-bf1e-e3c085129f33" Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.944320 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.944733 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.944575 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.944812 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.944842 5117 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.944900 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.944918 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.944930 5117 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.944958 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:01.944926585 +0000 UTC m=+95.238213797 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.944992 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:01.944977677 +0000 UTC m=+95.238264889 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.944778 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.945012 5117 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.945050 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:01.945032129 +0000 UTC m=+95.238319301 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:12:57 crc kubenswrapper[5117]: I0321 00:12:57.945078 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.945170 5117 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:12:57 crc kubenswrapper[5117]: E0321 00:12:57.945196 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:01.945188972 +0000 UTC m=+95.238476144 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:12:58 crc kubenswrapper[5117]: I0321 00:12:58.017704 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" event={"ID":"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc","Type":"ContainerStarted","Data":"e7d27724f8bbcb76f5471f968551232d957ddde8dff240b6c3e390cc72b8964d"} Mar 21 00:12:58 crc kubenswrapper[5117]: I0321 00:12:58.019704 5117 generic.go:358] "Generic (PLEG): container finished" podID="e685d70e-7b85-464d-b251-f682148f431f" containerID="9da9ecf69ebeb25146b645d2c60cec6840b1730edf09ecef607181a9dd25e8dc" exitCode=0 Mar 21 00:12:58 crc kubenswrapper[5117]: I0321 00:12:58.019771 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerDied","Data":"9da9ecf69ebeb25146b645d2c60cec6840b1730edf09ecef607181a9dd25e8dc"} Mar 21 00:12:58 crc kubenswrapper[5117]: I0321 00:12:58.020996 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" event={"ID":"d1770ea5-3738-4407-8de2-c4d3aea8385e","Type":"ContainerStarted","Data":"278ccf0226812fda11663b760ec0310f911c75f24c49189eed99f64a95a139e2"} Mar 21 00:12:58 crc kubenswrapper[5117]: I0321 00:12:58.046013 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:58 crc kubenswrapper[5117]: E0321 00:12:58.046173 5117 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:12:58 crc kubenswrapper[5117]: E0321 00:12:58.046238 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs podName:482a0770-6dc5-4682-bf1e-e3c085129f33 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:02.046222259 +0000 UTC m=+95.339509421 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs") pod "network-metrics-daemon-qp2t9" (UID: "482a0770-6dc5-4682-bf1e-e3c085129f33") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:12:58 crc kubenswrapper[5117]: I0321 00:12:58.054882 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-lmfdm" podStartSLOduration=70.054868253 podStartE2EDuration="1m10.054868253s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:12:58.054199583 +0000 UTC m=+91.347486765" watchObservedRunningTime="2026-03-21 00:12:58.054868253 +0000 UTC m=+91.348155425" Mar 21 00:12:58 crc kubenswrapper[5117]: I0321 00:12:58.146510 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:12:58 crc kubenswrapper[5117]: E0321 00:12:58.146666 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:02.146639167 +0000 UTC m=+95.439926339 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:12:59 crc kubenswrapper[5117]: I0321 00:12:59.025478 5117 generic.go:358] "Generic (PLEG): container finished" podID="e685d70e-7b85-464d-b251-f682148f431f" containerID="ed41497d05e257d9c9b1da7413b3a123931fdf5ce430b704d428362df4e5c12d" exitCode=0 Mar 21 00:12:59 crc kubenswrapper[5117]: I0321 00:12:59.025540 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerDied","Data":"ed41497d05e257d9c9b1da7413b3a123931fdf5ce430b704d428362df4e5c12d"} Mar 21 00:12:59 crc kubenswrapper[5117]: I0321 00:12:59.030562 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerStarted","Data":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} Mar 21 00:12:59 crc kubenswrapper[5117]: I0321 00:12:59.578604 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:12:59 crc kubenswrapper[5117]: E0321 00:12:59.579186 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 21 00:12:59 crc kubenswrapper[5117]: I0321 00:12:59.578676 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:12:59 crc kubenswrapper[5117]: E0321 00:12:59.579308 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 21 00:12:59 crc kubenswrapper[5117]: I0321 00:12:59.578706 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:12:59 crc kubenswrapper[5117]: E0321 00:12:59.579403 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qp2t9" podUID="482a0770-6dc5-4682-bf1e-e3c085129f33" Mar 21 00:12:59 crc kubenswrapper[5117]: I0321 00:12:59.578636 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:12:59 crc kubenswrapper[5117]: E0321 00:12:59.579488 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 21 00:13:00 crc kubenswrapper[5117]: I0321 00:13:00.036605 5117 generic.go:358] "Generic (PLEG): container finished" podID="e685d70e-7b85-464d-b251-f682148f431f" containerID="90cf2801e09be1541d9177a8e44ca8a4f84783938352ca5937fc2484eb30c2ec" exitCode=0 Mar 21 00:13:00 crc kubenswrapper[5117]: I0321 00:13:00.036725 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerDied","Data":"90cf2801e09be1541d9177a8e44ca8a4f84783938352ca5937fc2484eb30c2ec"} Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.046598 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerStarted","Data":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.047015 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.047034 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.059261 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerStarted","Data":"67357919d90933854d7b88c7df5573b2128b361238a66d90a950ec0d02559bb9"} Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.080744 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" podStartSLOduration=73.080725923 podStartE2EDuration="1m13.080725923s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:01.079439914 +0000 UTC m=+94.372727106" watchObservedRunningTime="2026-03-21 00:13:01.080725923 +0000 UTC m=+94.374013095" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.083984 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.579020 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.579116 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.579211 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.579116 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.579370 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qp2t9" podUID="482a0770-6dc5-4682-bf1e-e3c085129f33" Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.579422 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.579555 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.580048 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.990765 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.991228 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.991286 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:13:01 crc kubenswrapper[5117]: I0321 00:13:01.991325 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.991063 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.991482 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.991499 5117 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.991574 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.991553356 +0000 UTC m=+103.284840538 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.992040 5117 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.992085 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.992075362 +0000 UTC m=+103.285362544 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.992148 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.992161 5117 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.992171 5117 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.991443 5117 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.992199 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.992190376 +0000 UTC m=+103.285477558 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 21 00:13:01 crc kubenswrapper[5117]: E0321 00:13:01.992465 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.992415233 +0000 UTC m=+103.285702485 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 21 00:13:02 crc kubenswrapper[5117]: I0321 00:13:02.064156 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:13:02 crc kubenswrapper[5117]: I0321 00:13:02.092632 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:02 crc kubenswrapper[5117]: E0321 00:13:02.092851 5117 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:13:02 crc kubenswrapper[5117]: E0321 00:13:02.092931 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs podName:482a0770-6dc5-4682-bf1e-e3c085129f33 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.092909233 +0000 UTC m=+103.386196415 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs") pod "network-metrics-daemon-qp2t9" (UID: "482a0770-6dc5-4682-bf1e-e3c085129f33") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 21 00:13:02 crc kubenswrapper[5117]: I0321 00:13:02.109401 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:13:02 crc kubenswrapper[5117]: I0321 00:13:02.193849 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:02 crc kubenswrapper[5117]: E0321 00:13:02.194062 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.194022521 +0000 UTC m=+103.487309733 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:03 crc kubenswrapper[5117]: I0321 00:13:03.070714 5117 generic.go:358] "Generic (PLEG): container finished" podID="e685d70e-7b85-464d-b251-f682148f431f" containerID="67357919d90933854d7b88c7df5573b2128b361238a66d90a950ec0d02559bb9" exitCode=0 Mar 21 00:13:03 crc kubenswrapper[5117]: I0321 00:13:03.070809 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerDied","Data":"67357919d90933854d7b88c7df5573b2128b361238a66d90a950ec0d02559bb9"} Mar 21 00:13:03 crc kubenswrapper[5117]: I0321 00:13:03.579312 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:13:03 crc kubenswrapper[5117]: E0321 00:13:03.579492 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 21 00:13:03 crc kubenswrapper[5117]: I0321 00:13:03.579518 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:03 crc kubenswrapper[5117]: E0321 00:13:03.579666 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 21 00:13:03 crc kubenswrapper[5117]: I0321 00:13:03.579728 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:03 crc kubenswrapper[5117]: E0321 00:13:03.579786 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qp2t9" podUID="482a0770-6dc5-4682-bf1e-e3c085129f33" Mar 21 00:13:03 crc kubenswrapper[5117]: I0321 00:13:03.579834 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:03 crc kubenswrapper[5117]: E0321 00:13:03.579878 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 21 00:13:03 crc kubenswrapper[5117]: I0321 00:13:03.613392 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-qp2t9"] Mar 21 00:13:04 crc kubenswrapper[5117]: I0321 00:13:04.079378 5117 generic.go:358] "Generic (PLEG): container finished" podID="e685d70e-7b85-464d-b251-f682148f431f" containerID="9082bfd2c122dd09cc32fbd60b1bb2b17ecbc784fbd7eb84db25762c1e4c564c" exitCode=0 Mar 21 00:13:04 crc kubenswrapper[5117]: I0321 00:13:04.079507 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerDied","Data":"9082bfd2c122dd09cc32fbd60b1bb2b17ecbc784fbd7eb84db25762c1e4c564c"} Mar 21 00:13:04 crc kubenswrapper[5117]: I0321 00:13:04.080110 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:04 crc kubenswrapper[5117]: E0321 00:13:04.080288 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qp2t9" podUID="482a0770-6dc5-4682-bf1e-e3c085129f33" Mar 21 00:13:05 crc kubenswrapper[5117]: I0321 00:13:05.087214 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" event={"ID":"e685d70e-7b85-464d-b251-f682148f431f","Type":"ContainerStarted","Data":"71468d94d92592d7e22f704cab16d9c422a2915ee4ce24731ab50ca0e5ea8abe"} Mar 21 00:13:05 crc kubenswrapper[5117]: I0321 00:13:05.106804 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-gzrb6" podStartSLOduration=77.106781596 podStartE2EDuration="1m17.106781596s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:05.105725765 +0000 UTC m=+98.399012947" watchObservedRunningTime="2026-03-21 00:13:05.106781596 +0000 UTC m=+98.400068768" Mar 21 00:13:05 crc kubenswrapper[5117]: I0321 00:13:05.578980 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:05 crc kubenswrapper[5117]: I0321 00:13:05.579021 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:05 crc kubenswrapper[5117]: I0321 00:13:05.578982 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:13:05 crc kubenswrapper[5117]: E0321 00:13:05.579152 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 21 00:13:05 crc kubenswrapper[5117]: E0321 00:13:05.579404 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 21 00:13:05 crc kubenswrapper[5117]: I0321 00:13:05.579415 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:05 crc kubenswrapper[5117]: E0321 00:13:05.579550 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qp2t9" podUID="482a0770-6dc5-4682-bf1e-e3c085129f33" Mar 21 00:13:05 crc kubenswrapper[5117]: E0321 00:13:05.579754 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 21 00:13:05 crc kubenswrapper[5117]: I0321 00:13:05.580959 5117 scope.go:117] "RemoveContainer" containerID="79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17" Mar 21 00:13:05 crc kubenswrapper[5117]: E0321 00:13:05.581361 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.578855 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.578899 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.578951 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:13:07 crc kubenswrapper[5117]: E0321 00:13:07.581172 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Mar 21 00:13:07 crc kubenswrapper[5117]: E0321 00:13:07.581290 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qp2t9" podUID="482a0770-6dc5-4682-bf1e-e3c085129f33" Mar 21 00:13:07 crc kubenswrapper[5117]: E0321 00:13:07.581378 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.581435 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:07 crc kubenswrapper[5117]: E0321 00:13:07.581498 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.873789 5117 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeReady" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.874398 5117 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.954357 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-dx9lp"] Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.970923 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz"] Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.971198 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.974230 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-67c89758df-fj8qh"] Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.974465 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.976509 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-755bb95488-d9bzl"] Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.976806 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.991789 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"config\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.992126 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.992369 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-djmfg\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.992404 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.992545 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.992842 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-9ddfb9f55-nsfg8"] Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.993371 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.993416 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-kl6m8\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.993543 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.993627 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"serving-cert\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.993833 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-client\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.994233 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.997604 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.997667 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"audit-1\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.997779 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.997897 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-qqw4z\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.998061 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"serving-cert\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.998142 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.998193 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"client-ca\"" Mar 21 00:13:07 crc kubenswrapper[5117]: I0321 00:13:07.998136 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.001818 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.001841 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.001959 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.002127 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.012490 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.014818 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.034359 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-4zqgh\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.035558 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-images\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.035736 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.035788 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"audit-1\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.035866 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"etcd-serving-ca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.036131 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-6n5ln\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.036298 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.036424 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.036586 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.036728 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"config\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.036877 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-tls\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.037027 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"etcd-client\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.036301 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.037196 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"image-import-ca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.037140 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-rbac-proxy\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.037317 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-global-ca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.038999 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"encryption-config-1\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.039081 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.040045 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-5777786469-xqfxc"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.040204 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.051081 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-config\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.051641 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.052013 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-dockercfg-6c46w\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.053030 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"cluster-image-registry-operator-dockercfg-ntnd7\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.053399 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.059038 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-64d44f6ddf-dg2qw"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.060125 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-operator-tls\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.065543 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.066168 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-747b44746d-7zctx"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.066345 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081513 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d565d0a-934e-4b80-a621-3af94b7578a9-trusted-ca-bundle\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081517 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"trusted-ca-bundle\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081558 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4wzp\" (UniqueName: \"kubernetes.io/projected/828ac2ae-7b43-4714-9cb9-a86252d5deeb-kube-api-access-d4wzp\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081639 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d565d0a-934e-4b80-a621-3af94b7578a9-serving-cert\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081661 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c68f3876-1fa5-43d6-b8a4-b278f689aaca-config\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081738 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d565d0a-934e-4b80-a621-3af94b7578a9-etcd-client\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081766 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5vfp\" (UniqueName: \"kubernetes.io/projected/c68f3876-1fa5-43d6-b8a4-b278f689aaca-kube-api-access-p5vfp\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081873 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/828ac2ae-7b43-4714-9cb9-a86252d5deeb-machine-api-operator-tls\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081927 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7d565d0a-934e-4b80-a621-3af94b7578a9-audit-dir\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081949 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2lvc\" (UniqueName: \"kubernetes.io/projected/7d565d0a-934e-4b80-a621-3af94b7578a9-kube-api-access-f2lvc\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.081975 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-config\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082065 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5a8eb05-2bff-483b-b42f-03b818f9a058-serving-cert\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082150 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c68f3876-1fa5-43d6-b8a4-b278f689aaca-serving-cert\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082205 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r878v\" (UniqueName: \"kubernetes.io/projected/a5a8eb05-2bff-483b-b42f-03b818f9a058-kube-api-access-r878v\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082373 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c68f3876-1fa5-43d6-b8a4-b278f689aaca-trusted-ca\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082407 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7d565d0a-934e-4b80-a621-3af94b7578a9-etcd-serving-ca\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082437 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/828ac2ae-7b43-4714-9cb9-a86252d5deeb-images\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082491 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7d565d0a-934e-4b80-a621-3af94b7578a9-encryption-config\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082518 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828ac2ae-7b43-4714-9cb9-a86252d5deeb-config\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082605 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082741 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-proxy-ca-bundles\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082767 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a5a8eb05-2bff-483b-b42f-03b818f9a058-tmp\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082954 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-client-ca\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.082988 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7d565d0a-934e-4b80-a621-3af94b7578a9-audit-policies\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.085550 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.086009 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-747b44746d-7zctx" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.094490 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"config-operator-serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.115417 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.116030 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"openshift-config-operator-dockercfg-sjn6s\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.116466 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.116599 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.116770 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.117362 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.117882 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.119043 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.122913 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29567520-5stkm"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.123584 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.124295 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.124479 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.124766 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"default-dockercfg-mdwwj\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.124924 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.125308 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-8dkm8\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.128828 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mtrnq"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.129389 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.129893 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.130474 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-mmcpt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.130625 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.131554 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"config\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.131882 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.132284 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"serviceca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.132864 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.133185 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"pruner-dockercfg-rs58m\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.139916 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"client-ca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.143978 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.144276 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.151198 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-d2bf2\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.151231 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.151522 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.151802 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.151997 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.152319 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-router-certs\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.152713 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.152763 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.152797 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.152926 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.152952 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"audit\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.153011 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-session\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.155449 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.156333 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.156970 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.167078 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-rbac-proxy\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.167292 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-wzhvk\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.167752 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-config\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.168036 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.168124 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-tls\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.169336 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.172769 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.186656 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-ocp-branding-template\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.187123 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c68f3876-1fa5-43d6-b8a4-b278f689aaca-serving-cert\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.191407 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-r878v\" (UniqueName: \"kubernetes.io/projected/a5a8eb05-2bff-483b-b42f-03b818f9a058-kube-api-access-r878v\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.191466 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/beedce19-bd88-4e81-a204-99cc121cd861-image-registry-operator-tls\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.191537 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/60516002-0c51-45a3-99a7-1ced5970206e-encryption-config\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.191566 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c68f3876-1fa5-43d6-b8a4-b278f689aaca-trusted-ca\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.191587 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bed6c59c-20c9-41cd-92c8-6ca66de85fc8-serving-cert\") pod \"openshift-apiserver-operator-846cbfc458-5zkpw\" (UID: \"bed6c59c-20c9-41cd-92c8-6ca66de85fc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.191616 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7d565d0a-934e-4b80-a621-3af94b7578a9-etcd-serving-ca\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.191639 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/828ac2ae-7b43-4714-9cb9-a86252d5deeb-images\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.191663 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beedce19-bd88-4e81-a204-99cc121cd861-trusted-ca\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.192806 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7d565d0a-934e-4b80-a621-3af94b7578a9-etcd-serving-ca\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.193449 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/828ac2ae-7b43-4714-9cb9-a86252d5deeb-images\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.193693 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c68f3876-1fa5-43d6-b8a4-b278f689aaca-trusted-ca\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.194072 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdnrf\" (UniqueName: \"kubernetes.io/projected/bed6c59c-20c9-41cd-92c8-6ca66de85fc8-kube-api-access-bdnrf\") pod \"openshift-apiserver-operator-846cbfc458-5zkpw\" (UID: \"bed6c59c-20c9-41cd-92c8-6ca66de85fc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.194100 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-config\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.194149 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7d565d0a-934e-4b80-a621-3af94b7578a9-encryption-config\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.194173 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828ac2ae-7b43-4714-9cb9-a86252d5deeb-config\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.194193 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-etcd-serving-ca\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.194216 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6498\" (UniqueName: \"kubernetes.io/projected/60516002-0c51-45a3-99a7-1ced5970206e-kube-api-access-c6498\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.194430 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-proxy-ca-bundles\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.194624 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a5a8eb05-2bff-483b-b42f-03b818f9a058-tmp\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.194652 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f6e0dec-9f57-403b-893c-5a30c576a799-serving-cert\") pod \"openshift-config-operator-5777786469-xqfxc\" (UID: \"9f6e0dec-9f57-403b-893c-5a30c576a799\") " pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.195038 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/828ac2ae-7b43-4714-9cb9-a86252d5deeb-config\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.195984 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-client-ca\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196030 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-oauth-serving-cert\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196157 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7d565d0a-934e-4b80-a621-3af94b7578a9-audit-policies\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196289 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/beedce19-bd88-4e81-a204-99cc121cd861-ca-trust-extracted-pem\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196367 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d565d0a-934e-4b80-a621-3af94b7578a9-trusted-ca-bundle\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196406 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d4wzp\" (UniqueName: \"kubernetes.io/projected/828ac2ae-7b43-4714-9cb9-a86252d5deeb-kube-api-access-d4wzp\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196441 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc67s\" (UniqueName: \"kubernetes.io/projected/fda377c1-1e51-4241-a939-a44cc536104d-kube-api-access-gc67s\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196468 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60516002-0c51-45a3-99a7-1ced5970206e-serving-cert\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196495 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60516002-0c51-45a3-99a7-1ced5970206e-audit-dir\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196535 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d565d0a-934e-4b80-a621-3af94b7578a9-serving-cert\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196564 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6zf5\" (UniqueName: \"kubernetes.io/projected/beedce19-bd88-4e81-a204-99cc121cd861-kube-api-access-c6zf5\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196598 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/beedce19-bd88-4e81-a204-99cc121cd861-bound-sa-token\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196624 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-service-ca\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196677 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c68f3876-1fa5-43d6-b8a4-b278f689aaca-config\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196706 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d565d0a-934e-4b80-a621-3af94b7578a9-etcd-client\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196734 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/60516002-0c51-45a3-99a7-1ced5970206e-node-pullsecrets\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196764 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9f6e0dec-9f57-403b-893c-5a30c576a799-available-featuregates\") pod \"openshift-config-operator-5777786469-xqfxc\" (UID: \"9f6e0dec-9f57-403b-893c-5a30c576a799\") " pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196805 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p5vfp\" (UniqueName: \"kubernetes.io/projected/c68f3876-1fa5-43d6-b8a4-b278f689aaca-kube-api-access-p5vfp\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196834 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bed6c59c-20c9-41cd-92c8-6ca66de85fc8-config\") pod \"openshift-apiserver-operator-846cbfc458-5zkpw\" (UID: \"bed6c59c-20c9-41cd-92c8-6ca66de85fc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196875 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj8h7\" (UniqueName: \"kubernetes.io/projected/9f6e0dec-9f57-403b-893c-5a30c576a799-kube-api-access-fj8h7\") pod \"openshift-config-operator-5777786469-xqfxc\" (UID: \"9f6e0dec-9f57-403b-893c-5a30c576a799\") " pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196901 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7d565d0a-934e-4b80-a621-3af94b7578a9-audit-policies\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.196805 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-proxy-ca-bundles\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.197945 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d565d0a-934e-4b80-a621-3af94b7578a9-trusted-ca-bundle\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.197946 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/828ac2ae-7b43-4714-9cb9-a86252d5deeb-machine-api-operator-tls\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.198004 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c68f3876-1fa5-43d6-b8a4-b278f689aaca-config\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.199282 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-client-ca\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.199815 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-trusted-ca-bundle\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.199848 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-audit\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.199898 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60516002-0c51-45a3-99a7-1ced5970206e-etcd-client\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.199979 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7d565d0a-934e-4b80-a621-3af94b7578a9-audit-dir\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200010 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fda377c1-1e51-4241-a939-a44cc536104d-console-oauth-config\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200049 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7d565d0a-934e-4b80-a621-3af94b7578a9-audit-dir\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200082 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f2lvc\" (UniqueName: \"kubernetes.io/projected/7d565d0a-934e-4b80-a621-3af94b7578a9-kube-api-access-f2lvc\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200109 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-trusted-ca-bundle\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200308 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-config\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200338 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-console-config\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200730 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/beedce19-bd88-4e81-a204-99cc121cd861-tmp\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200790 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkjxk\" (UniqueName: \"kubernetes.io/projected/6c43692a-55c8-4d6a-bd36-9c3268722421-kube-api-access-vkjxk\") pod \"downloads-747b44746d-7zctx\" (UID: \"6c43692a-55c8-4d6a-bd36-9c3268722421\") " pod="openshift-console/downloads-747b44746d-7zctx" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200817 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fda377c1-1e51-4241-a939-a44cc536104d-console-serving-cert\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200860 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-image-import-ca\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.200898 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5a8eb05-2bff-483b-b42f-03b818f9a058-serving-cert\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.209615 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7d565d0a-934e-4b80-a621-3af94b7578a9-encryption-config\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.209703 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d565d0a-934e-4b80-a621-3af94b7578a9-etcd-client\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.209731 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d565d0a-934e-4b80-a621-3af94b7578a9-serving-cert\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.210270 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a5a8eb05-2bff-483b-b42f-03b818f9a058-tmp\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.211050 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/828ac2ae-7b43-4714-9cb9-a86252d5deeb-machine-api-operator-tls\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.213390 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5vfp\" (UniqueName: \"kubernetes.io/projected/c68f3876-1fa5-43d6-b8a4-b278f689aaca-kube-api-access-p5vfp\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.213414 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4wzp\" (UniqueName: \"kubernetes.io/projected/828ac2ae-7b43-4714-9cb9-a86252d5deeb-kube-api-access-d4wzp\") pod \"machine-api-operator-755bb95488-d9bzl\" (UID: \"828ac2ae-7b43-4714-9cb9-a86252d5deeb\") " pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.214096 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-config\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.214268 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-r878v\" (UniqueName: \"kubernetes.io/projected/a5a8eb05-2bff-483b-b42f-03b818f9a058-kube-api-access-r878v\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.214599 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-gjth4"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.215084 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.218416 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.218519 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.219165 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5a8eb05-2bff-483b-b42f-03b818f9a058-serving-cert\") pod \"controller-manager-65b6cccf98-dx9lp\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.219994 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-jmhxf\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.220089 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.220388 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2lvc\" (UniqueName: \"kubernetes.io/projected/7d565d0a-934e-4b80-a621-3af94b7578a9-kube-api-access-f2lvc\") pod \"apiserver-8596bd845d-nd9tz\" (UID: \"7d565d0a-934e-4b80-a621-3af94b7578a9\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.226453 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c68f3876-1fa5-43d6-b8a4-b278f689aaca-serving-cert\") pod \"console-operator-67c89758df-fj8qh\" (UID: \"c68f3876-1fa5-43d6-b8a4-b278f689aaca\") " pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.232896 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.247234 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.247413 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.248146 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.274293 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-dockercfg-bf7fj\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.289466 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.293973 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.297045 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302012 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/60516002-0c51-45a3-99a7-1ced5970206e-node-pullsecrets\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302052 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9f6e0dec-9f57-403b-893c-5a30c576a799-available-featuregates\") pod \"openshift-config-operator-5777786469-xqfxc\" (UID: \"9f6e0dec-9f57-403b-893c-5a30c576a799\") " pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302080 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302100 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29e146a-be7b-49f8-b97b-31f4c976afb5-serving-cert\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302118 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bed6c59c-20c9-41cd-92c8-6ca66de85fc8-config\") pod \"openshift-apiserver-operator-846cbfc458-5zkpw\" (UID: \"bed6c59c-20c9-41cd-92c8-6ca66de85fc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302136 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fj8h7\" (UniqueName: \"kubernetes.io/projected/9f6e0dec-9f57-403b-893c-5a30c576a799-kube-api-access-fj8h7\") pod \"openshift-config-operator-5777786469-xqfxc\" (UID: \"9f6e0dec-9f57-403b-893c-5a30c576a799\") " pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302151 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a29e146a-be7b-49f8-b97b-31f4c976afb5-tmp\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302170 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a9dda43-9285-4470-8175-f919dd943c17-samples-operator-tls\") pod \"cluster-samples-operator-6b564684c8-wzx89\" (UID: \"3a9dda43-9285-4470-8175-f919dd943c17\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302217 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-trusted-ca-bundle\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302234 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-audit\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302269 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302294 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-error\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302314 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aed5803f-cd47-4e06-b5bf-059b2d51e70d-auth-proxy-config\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302334 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60516002-0c51-45a3-99a7-1ced5970206e-etcd-client\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302353 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-login\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302372 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302400 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fda377c1-1e51-4241-a939-a44cc536104d-console-oauth-config\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302421 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-trusted-ca-bundle\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302441 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-client-ca\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302472 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-console-config\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302493 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb9gs\" (UniqueName: \"kubernetes.io/projected/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-kube-api-access-sb9gs\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302511 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pksg4\" (UniqueName: \"kubernetes.io/projected/a29e146a-be7b-49f8-b97b-31f4c976afb5-kube-api-access-pksg4\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302532 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/beedce19-bd88-4e81-a204-99cc121cd861-tmp\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302550 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302567 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed5803f-cd47-4e06-b5bf-059b2d51e70d-config\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302587 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vkjxk\" (UniqueName: \"kubernetes.io/projected/6c43692a-55c8-4d6a-bd36-9c3268722421-kube-api-access-vkjxk\") pod \"downloads-747b44746d-7zctx\" (UID: \"6c43692a-55c8-4d6a-bd36-9c3268722421\") " pod="openshift-console/downloads-747b44746d-7zctx" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302605 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302624 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fda377c1-1e51-4241-a939-a44cc536104d-console-serving-cert\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302639 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-image-import-ca\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302658 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkwnz\" (UniqueName: \"kubernetes.io/projected/79fcf4ae-c05f-492c-88f0-ec441eb255aa-kube-api-access-tkwnz\") pod \"image-pruner-29567520-5stkm\" (UID: \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\") " pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302679 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sw5t\" (UniqueName: \"kubernetes.io/projected/aed5803f-cd47-4e06-b5bf-059b2d51e70d-kube-api-access-6sw5t\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302696 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62xzd\" (UniqueName: \"kubernetes.io/projected/3a9dda43-9285-4470-8175-f919dd943c17-kube-api-access-62xzd\") pod \"cluster-samples-operator-6b564684c8-wzx89\" (UID: \"3a9dda43-9285-4470-8175-f919dd943c17\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302716 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/beedce19-bd88-4e81-a204-99cc121cd861-image-registry-operator-tls\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302739 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/79fcf4ae-c05f-492c-88f0-ec441eb255aa-serviceca\") pod \"image-pruner-29567520-5stkm\" (UID: \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\") " pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302758 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-dir\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302776 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-policies\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302791 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-session\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302807 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-service-ca\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302836 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/60516002-0c51-45a3-99a7-1ced5970206e-encryption-config\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302852 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-router-certs\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302870 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bed6c59c-20c9-41cd-92c8-6ca66de85fc8-serving-cert\") pod \"openshift-apiserver-operator-846cbfc458-5zkpw\" (UID: \"bed6c59c-20c9-41cd-92c8-6ca66de85fc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302891 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beedce19-bd88-4e81-a204-99cc121cd861-trusted-ca\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302909 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bdnrf\" (UniqueName: \"kubernetes.io/projected/bed6c59c-20c9-41cd-92c8-6ca66de85fc8-kube-api-access-bdnrf\") pod \"openshift-apiserver-operator-846cbfc458-5zkpw\" (UID: \"bed6c59c-20c9-41cd-92c8-6ca66de85fc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302924 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-config\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302954 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-etcd-serving-ca\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302972 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-c6498\" (UniqueName: \"kubernetes.io/projected/60516002-0c51-45a3-99a7-1ced5970206e-kube-api-access-c6498\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.302994 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f6e0dec-9f57-403b-893c-5a30c576a799-serving-cert\") pod \"openshift-config-operator-5777786469-xqfxc\" (UID: \"9f6e0dec-9f57-403b-893c-5a30c576a799\") " pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303015 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-oauth-serving-cert\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303039 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/aed5803f-cd47-4e06-b5bf-059b2d51e70d-machine-approver-tls\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303060 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/beedce19-bd88-4e81-a204-99cc121cd861-ca-trust-extracted-pem\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303088 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gc67s\" (UniqueName: \"kubernetes.io/projected/fda377c1-1e51-4241-a939-a44cc536104d-kube-api-access-gc67s\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303105 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60516002-0c51-45a3-99a7-1ced5970206e-serving-cert\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303122 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60516002-0c51-45a3-99a7-1ced5970206e-audit-dir\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303141 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-config\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303165 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-c6zf5\" (UniqueName: \"kubernetes.io/projected/beedce19-bd88-4e81-a204-99cc121cd861-kube-api-access-c6zf5\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303185 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/beedce19-bd88-4e81-a204-99cc121cd861-bound-sa-token\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303200 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-service-ca\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303218 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.303354 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/60516002-0c51-45a3-99a7-1ced5970206e-node-pullsecrets\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.304550 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9f6e0dec-9f57-403b-893c-5a30c576a799-available-featuregates\") pod \"openshift-config-operator-5777786469-xqfxc\" (UID: \"9f6e0dec-9f57-403b-893c-5a30c576a799\") " pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.304710 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60516002-0c51-45a3-99a7-1ced5970206e-audit-dir\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.305334 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-audit\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.305616 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-trusted-ca-bundle\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.305870 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-trusted-ca-bundle\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.306072 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-console-config\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.306566 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bed6c59c-20c9-41cd-92c8-6ca66de85fc8-config\") pod \"openshift-apiserver-operator-846cbfc458-5zkpw\" (UID: \"bed6c59c-20c9-41cd-92c8-6ca66de85fc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.307135 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-etcd-serving-ca\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.307492 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beedce19-bd88-4e81-a204-99cc121cd861-trusted-ca\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.308101 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/beedce19-bd88-4e81-a204-99cc121cd861-ca-trust-extracted-pem\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.308149 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/beedce19-bd88-4e81-a204-99cc121cd861-tmp\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.308597 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-image-import-ca\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.308879 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-service-ca\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.308971 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fda377c1-1e51-4241-a939-a44cc536104d-oauth-serving-cert\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.309109 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60516002-0c51-45a3-99a7-1ced5970206e-config\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.310098 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/beedce19-bd88-4e81-a204-99cc121cd861-image-registry-operator-tls\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.311623 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bed6c59c-20c9-41cd-92c8-6ca66de85fc8-serving-cert\") pod \"openshift-apiserver-operator-846cbfc458-5zkpw\" (UID: \"bed6c59c-20c9-41cd-92c8-6ca66de85fc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.311639 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f6e0dec-9f57-403b-893c-5a30c576a799-serving-cert\") pod \"openshift-config-operator-5777786469-xqfxc\" (UID: \"9f6e0dec-9f57-403b-893c-5a30c576a799\") " pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.311789 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60516002-0c51-45a3-99a7-1ced5970206e-serving-cert\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.312618 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/60516002-0c51-45a3-99a7-1ced5970206e-etcd-client\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.313728 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/60516002-0c51-45a3-99a7-1ced5970206e-encryption-config\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.314018 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.314771 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fda377c1-1e51-4241-a939-a44cc536104d-console-oauth-config\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.315938 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fda377c1-1e51-4241-a939-a44cc536104d-console-serving-cert\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.331257 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-799b87ffcd-p2jqd"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.331522 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.333302 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.333729 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-config\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.351624 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.355037 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.381346 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.398646 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-6w67b\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404659 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-router-certs\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404735 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/aed5803f-cd47-4e06-b5bf-059b2d51e70d-machine-approver-tls\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404768 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-config\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404796 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404826 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404872 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29e146a-be7b-49f8-b97b-31f4c976afb5-serving-cert\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404897 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a29e146a-be7b-49f8-b97b-31f4c976afb5-tmp\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404923 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a9dda43-9285-4470-8175-f919dd943c17-samples-operator-tls\") pod \"cluster-samples-operator-6b564684c8-wzx89\" (UID: \"3a9dda43-9285-4470-8175-f919dd943c17\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404973 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.404998 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-error\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405021 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aed5803f-cd47-4e06-b5bf-059b2d51e70d-auth-proxy-config\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405049 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-login\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405071 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405109 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-client-ca\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405132 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sb9gs\" (UniqueName: \"kubernetes.io/projected/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-kube-api-access-sb9gs\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405155 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pksg4\" (UniqueName: \"kubernetes.io/projected/a29e146a-be7b-49f8-b97b-31f4c976afb5-kube-api-access-pksg4\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405180 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405209 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed5803f-cd47-4e06-b5bf-059b2d51e70d-config\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405235 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405296 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-tkwnz\" (UniqueName: \"kubernetes.io/projected/79fcf4ae-c05f-492c-88f0-ec441eb255aa-kube-api-access-tkwnz\") pod \"image-pruner-29567520-5stkm\" (UID: \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\") " pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405345 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6sw5t\" (UniqueName: \"kubernetes.io/projected/aed5803f-cd47-4e06-b5bf-059b2d51e70d-kube-api-access-6sw5t\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405368 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-62xzd\" (UniqueName: \"kubernetes.io/projected/3a9dda43-9285-4470-8175-f919dd943c17-kube-api-access-62xzd\") pod \"cluster-samples-operator-6b564684c8-wzx89\" (UID: \"3a9dda43-9285-4470-8175-f919dd943c17\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405391 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/79fcf4ae-c05f-492c-88f0-ec441eb255aa-serviceca\") pod \"image-pruner-29567520-5stkm\" (UID: \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\") " pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405411 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-dir\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405435 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-policies\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405459 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-session\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.405493 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-service-ca\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.406542 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-service-ca\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.407235 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed5803f-cd47-4e06-b5bf-059b2d51e70d-config\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.408644 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-client-ca\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.408714 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-dir\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.408962 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.409947 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-policies\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.410615 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/79fcf4ae-c05f-492c-88f0-ec441eb255aa-serviceca\") pod \"image-pruner-29567520-5stkm\" (UID: \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\") " pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.410951 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aed5803f-cd47-4e06-b5bf-059b2d51e70d-auth-proxy-config\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.411207 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a29e146a-be7b-49f8-b97b-31f4c976afb5-tmp\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.412437 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-config\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.416817 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.417879 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-session\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.418533 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-login\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.418744 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.421942 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29e146a-be7b-49f8-b97b-31f4c976afb5-serving-cert\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.423895 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.423882 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-router-certs\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.426798 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a9dda43-9285-4470-8175-f919dd943c17-samples-operator-tls\") pod \"cluster-samples-operator-6b564684c8-wzx89\" (UID: \"3a9dda43-9285-4470-8175-f919dd943c17\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.427728 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-67c89758df-fj8qh"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.427775 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-5777786469-xqfxc"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.427915 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-error\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.427941 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.428615 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-dx9lp"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.428640 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.428660 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.428677 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.428709 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.429900 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.433715 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/aed5803f-cd47-4e06-b5bf-059b2d51e70d-machine-approver-tls\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.461014 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj8h7\" (UniqueName: \"kubernetes.io/projected/9f6e0dec-9f57-403b-893c-5a30c576a799-kube-api-access-fj8h7\") pod \"openshift-config-operator-5777786469-xqfxc\" (UID: \"9f6e0dec-9f57-403b-893c-5a30c576a799\") " pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.479806 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6498\" (UniqueName: \"kubernetes.io/projected/60516002-0c51-45a3-99a7-1ced5970206e-kube-api-access-c6498\") pod \"apiserver-9ddfb9f55-nsfg8\" (UID: \"60516002-0c51-45a3-99a7-1ced5970206e\") " pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.493572 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdnrf\" (UniqueName: \"kubernetes.io/projected/bed6c59c-20c9-41cd-92c8-6ca66de85fc8-kube-api-access-bdnrf\") pod \"openshift-apiserver-operator-846cbfc458-5zkpw\" (UID: \"bed6c59c-20c9-41cd-92c8-6ca66de85fc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.495228 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.524030 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkjxk\" (UniqueName: \"kubernetes.io/projected/6c43692a-55c8-4d6a-bd36-9c3268722421-kube-api-access-vkjxk\") pod \"downloads-747b44746d-7zctx\" (UID: \"6c43692a-55c8-4d6a-bd36-9c3268722421\") " pod="openshift-console/downloads-747b44746d-7zctx" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.532474 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc67s\" (UniqueName: \"kubernetes.io/projected/fda377c1-1e51-4241-a939-a44cc536104d-kube-api-access-gc67s\") pod \"console-64d44f6ddf-dg2qw\" (UID: \"fda377c1-1e51-4241-a939-a44cc536104d\") " pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.547416 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6zf5\" (UniqueName: \"kubernetes.io/projected/beedce19-bd88-4e81-a204-99cc121cd861-kube-api-access-c6zf5\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.569466 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/beedce19-bd88-4e81-a204-99cc121cd861-bound-sa-token\") pod \"cluster-image-registry-operator-86c45576b9-89dmq\" (UID: \"beedce19-bd88-4e81-a204-99cc121cd861\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.594054 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-6tbpn\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.614135 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"serving-cert\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.625754 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.626397 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.626947 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.634173 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"service-ca-bundle\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.643223 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.643790 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:08 crc kubenswrapper[5117]: W0321 00:13:08.652573 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5a8eb05_2bff_483b_b42f_03b818f9a058.slice/crio-df46ff372feb0a56a5e8d86f9ba8582bbe49ad744382c4df90099fdf1892bda8 WatchSource:0}: Error finding container df46ff372feb0a56a5e8d86f9ba8582bbe49ad744382c4df90099fdf1892bda8: Status 404 returned error can't find the container with id df46ff372feb0a56a5e8d86f9ba8582bbe49ad744382c4df90099fdf1892bda8 Mar 21 00:13:08 crc kubenswrapper[5117]: W0321 00:13:08.657490 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d565d0a_934e_4b80_a621_3af94b7578a9.slice/crio-4c1fbcf7647d2908d04bf3fb6e8c66dd481abdbfb87e0c3cf43e516d1a0a694d WatchSource:0}: Error finding container 4c1fbcf7647d2908d04bf3fb6e8c66dd481abdbfb87e0c3cf43e516d1a0a694d: Status 404 returned error can't find the container with id 4c1fbcf7647d2908d04bf3fb6e8c66dd481abdbfb87e0c3cf43e516d1a0a694d Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.659349 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"trusted-ca-bundle\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.668925 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.669142 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.673387 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.676434 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.677613 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.680004 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.680215 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.680550 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.680556 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.694832 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-68cf44c8b8-zkpbt"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.696068 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.709688 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.709825 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.709860 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.709884 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.713218 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-config\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.718194 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.728668 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.728935 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.729003 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2n4q9"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.734543 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.743443 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.750388 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-747b44746d-7zctx"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.750463 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.750614 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.756125 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.757479 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-62xzd\" (UniqueName: \"kubernetes.io/projected/3a9dda43-9285-4470-8175-f919dd943c17-kube-api-access-62xzd\") pod \"cluster-samples-operator-6b564684c8-wzx89\" (UID: \"3a9dda43-9285-4470-8175-f919dd943c17\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.759381 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.759663 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.764322 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-69db94689b-twxrl"] Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.773388 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-747b44746d-7zctx" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.785264 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb9gs\" (UniqueName: \"kubernetes.io/projected/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-kube-api-access-sb9gs\") pod \"oauth-openshift-66458b6674-mtrnq\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.796102 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pksg4\" (UniqueName: \"kubernetes.io/projected/a29e146a-be7b-49f8-b97b-31f4c976afb5-kube-api-access-pksg4\") pod \"route-controller-manager-776cdc94d6-l6ptk\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.801447 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.811267 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkwnz\" (UniqueName: \"kubernetes.io/projected/79fcf4ae-c05f-492c-88f0-ec441eb255aa-kube-api-access-tkwnz\") pod \"image-pruner-29567520-5stkm\" (UID: \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\") " pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.839052 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sw5t\" (UniqueName: \"kubernetes.io/projected/aed5803f-cd47-4e06-b5bf-059b2d51e70d-kube-api-access-6sw5t\") pod \"machine-approver-54c688565-wrfjw\" (UID: \"aed5803f-cd47-4e06-b5bf-059b2d51e70d\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.843911 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.850685 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.856577 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"metrics-tls\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.874698 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"dns-operator-dockercfg-wbbsn\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.894504 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.915543 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.970117 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-dockercfg-jcmfj\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.973365 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:08 crc kubenswrapper[5117]: I0321 00:13:08.994127 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-config\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.013786 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2677t\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-kube-api-access-2677t\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.013828 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b91d27f8-a52a-4a5a-a913-782a11c8e051-kube-api-access\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.013866 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-bound-sa-token\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.013911 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f85237c3-327f-4a9e-8496-00c9087e5aa7-installation-pull-secrets\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.013930 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-certificates\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.013946 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-trusted-ca\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.013981 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f85237c3-327f-4a9e-8496-00c9087e5aa7-ca-trust-extracted\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.013995 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91d27f8-a52a-4a5a-a913-782a11c8e051-config\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.014032 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-tls\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.014054 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.014071 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b91d27f8-a52a-4a5a-a913-782a11c8e051-serving-cert\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.014110 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/b91d27f8-a52a-4a5a-a913-782a11c8e051-tmp-dir\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.014534 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.514518816 +0000 UTC m=+102.807805988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.015488 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-serving-cert\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.017579 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.017991 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.018079 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.033905 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.059087 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: W0321 00:13:09.069638 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbed6c59c_20c9_41cd_92c8_6ca66de85fc8.slice/crio-f8f4acef8cb2dbadfef57969c6aa39ea20cb0fe456870928c4a73a7c812035d5 WatchSource:0}: Error finding container f8f4acef8cb2dbadfef57969c6aa39ea20cb0fe456870928c4a73a7c812035d5: Status 404 returned error can't find the container with id f8f4acef8cb2dbadfef57969c6aa39ea20cb0fe456870928c4a73a7c812035d5 Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.074461 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-serving-cert\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.088558 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.090421 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.095129 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-dockercfg-tnfx9\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.114825 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115029 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bd48a05d-1bac-41bb-8b91-419cb4072895-profile-collector-cert\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115061 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-trusted-ca\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115086 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9e554a8-6f94-4a46-acd7-3446e17421a5-config\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115115 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c5113992-e2f4-427c-acf3-55a4ef62376c-tmp-dir\") pod \"dns-operator-799b87ffcd-p2jqd\" (UID: \"c5113992-e2f4-427c-acf3-55a4ef62376c\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115135 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-etcd-client\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115154 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ed37e18-269e-4b50-a36a-0a723d5377d7-kube-api-access\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115171 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bntmk\" (UniqueName: \"kubernetes.io/projected/c5113992-e2f4-427c-acf3-55a4ef62376c-kube-api-access-bntmk\") pod \"dns-operator-799b87ffcd-p2jqd\" (UID: \"c5113992-e2f4-427c-acf3-55a4ef62376c\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115190 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl6ml\" (UniqueName: \"kubernetes.io/projected/1a398b4e-beca-4bbc-a41a-76f738f128d8-kube-api-access-cl6ml\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115210 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8ab52c-cb4f-467e-9044-1772b203fea8-config\") pod \"kube-storage-version-migrator-operator-565b79b866-x4xb5\" (UID: \"9d8ab52c-cb4f-467e-9044-1772b203fea8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115238 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bd48a05d-1bac-41bb-8b91-419cb4072895-srv-cert\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115270 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-mountpoint-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115287 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/35de0468-bf61-4986-8473-77b7de39194a-tmp\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115306 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b91d27f8-a52a-4a5a-a913-782a11c8e051-kube-api-access\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115335 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-etcd-ca\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115370 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-registration-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115395 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9e554a8-6f94-4a46-acd7-3446e17421a5-trusted-ca-bundle\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115416 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsb42\" (UniqueName: \"kubernetes.io/projected/9462f609-9733-4abc-8f7c-f5fa0041f39d-kube-api-access-qsb42\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115439 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prdxn\" (UniqueName: \"kubernetes.io/projected/55cf9bde-83fe-4510-a1ae-acaf82c35b62-kube-api-access-prdxn\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115488 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a398b4e-beca-4bbc-a41a-76f738f128d8-bound-sa-token\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115509 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ed37e18-269e-4b50-a36a-0a723d5377d7-config\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115540 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-plugins-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115559 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/58bdc39b-9471-4649-b1c4-45206dd8855d-tmp\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115578 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94zxc\" (UniqueName: \"kubernetes.io/projected/1280d6d3-8430-41a6-9e76-5ddb21a0f44e-kube-api-access-94zxc\") pod \"machine-config-controller-f9cdd68f7-cpsp6\" (UID: \"1280d6d3-8430-41a6-9e76-5ddb21a0f44e\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115606 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5113992-e2f4-427c-acf3-55a4ef62376c-metrics-tls\") pod \"dns-operator-799b87ffcd-p2jqd\" (UID: \"c5113992-e2f4-427c-acf3-55a4ef62376c\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115624 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-serving-cert\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115649 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2677t\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-kube-api-access-2677t\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115667 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/7ed37e18-269e-4b50-a36a-0a723d5377d7-tmp-dir\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115684 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1280d6d3-8430-41a6-9e76-5ddb21a0f44e-mcc-auth-proxy-config\") pod \"machine-config-controller-f9cdd68f7-cpsp6\" (UID: \"1280d6d3-8430-41a6-9e76-5ddb21a0f44e\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115701 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55cf9bde-83fe-4510-a1ae-acaf82c35b62-service-ca-bundle\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115708 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-config\"" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.115820 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.615801609 +0000 UTC m=+102.909088781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.115722 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a398b4e-beca-4bbc-a41a-76f738f128d8-trusted-ca\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116037 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9e554a8-6f94-4a46-acd7-3446e17421a5-serving-cert\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116069 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-config\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116107 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bd48a05d-1bac-41bb-8b91-419cb4072895-tmpfs\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116132 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4zg6\" (UniqueName: \"kubernetes.io/projected/9d8ab52c-cb4f-467e-9044-1772b203fea8-kube-api-access-d4zg6\") pod \"kube-storage-version-migrator-operator-565b79b866-x4xb5\" (UID: \"9d8ab52c-cb4f-467e-9044-1772b203fea8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116168 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-bound-sa-token\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116188 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a398b4e-beca-4bbc-a41a-76f738f128d8-metrics-tls\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116243 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/16a553dc-4b35-4c23-877a-4c7915ad94a4-apiservice-cert\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116287 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f85237c3-327f-4a9e-8496-00c9087e5aa7-installation-pull-secrets\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116310 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58bdc39b-9471-4649-b1c4-45206dd8855d-config\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116333 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/58bdc39b-9471-4649-b1c4-45206dd8855d-kube-api-access\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116364 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-certificates\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116393 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz6bd\" (UniqueName: \"kubernetes.io/projected/bd48a05d-1bac-41bb-8b91-419cb4072895-kube-api-access-fz6bd\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116420 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35de0468-bf61-4986-8473-77b7de39194a-serving-cert\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116467 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/70789848-de09-49a1-8f00-e29b310194fa-webhook-certs\") pod \"multus-admission-controller-69db94689b-twxrl\" (UID: \"70789848-de09-49a1-8f00-e29b310194fa\") " pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116492 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55cf9bde-83fe-4510-a1ae-acaf82c35b62-metrics-certs\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116515 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f85237c3-327f-4a9e-8496-00c9087e5aa7-ca-trust-extracted\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116534 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91d27f8-a52a-4a5a-a913-782a11c8e051-config\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116559 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1280d6d3-8430-41a6-9e76-5ddb21a0f44e-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-cpsp6\" (UID: \"1280d6d3-8430-41a6-9e76-5ddb21a0f44e\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116579 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/16a553dc-4b35-4c23-877a-4c7915ad94a4-webhook-cert\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116602 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35de0468-bf61-4986-8473-77b7de39194a-config\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116619 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58bdc39b-9471-4649-b1c4-45206dd8855d-serving-cert\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116653 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x7pt\" (UniqueName: \"kubernetes.io/projected/d9e554a8-6f94-4a46-acd7-3446e17421a5-kube-api-access-6x7pt\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116688 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-secret-volume\") pod \"collect-profiles-29567520-pnrmd\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116707 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-socket-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116737 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116755 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/b91d27f8-a52a-4a5a-a913-782a11c8e051-tmp-dir\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116776 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-tmp-dir\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116807 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-csi-data-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116828 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ed37e18-269e-4b50-a36a-0a723d5377d7-serving-cert\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116845 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-tls\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116862 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d5tr\" (UniqueName: \"kubernetes.io/projected/70789848-de09-49a1-8f00-e29b310194fa-kube-api-access-6d5tr\") pod \"multus-admission-controller-69db94689b-twxrl\" (UID: \"70789848-de09-49a1-8f00-e29b310194fa\") " pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116878 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/16a553dc-4b35-4c23-877a-4c7915ad94a4-tmpfs\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116895 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b91d27f8-a52a-4a5a-a913-782a11c8e051-serving-cert\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116916 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9f2t\" (UniqueName: \"kubernetes.io/projected/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-kube-api-access-b9f2t\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116940 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvhtm\" (UniqueName: \"kubernetes.io/projected/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-kube-api-access-pvhtm\") pod \"collect-profiles-29567520-pnrmd\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116958 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndrkw\" (UniqueName: \"kubernetes.io/projected/35de0468-bf61-4986-8473-77b7de39194a-kube-api-access-ndrkw\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.116977 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-etcd-service-ca\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.117008 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9e554a8-6f94-4a46-acd7-3446e17421a5-service-ca-bundle\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.117027 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9tc4\" (UniqueName: \"kubernetes.io/projected/16a553dc-4b35-4c23-877a-4c7915ad94a4-kube-api-access-f9tc4\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.117092 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-config-volume\") pod \"collect-profiles-29567520-pnrmd\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.117109 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/55cf9bde-83fe-4510-a1ae-acaf82c35b62-default-certificate\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.117120 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-trusted-ca\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.117128 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/55cf9bde-83fe-4510-a1ae-acaf82c35b62-stats-auth\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.117172 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d8ab52c-cb4f-467e-9044-1772b203fea8-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-x4xb5\" (UID: \"9d8ab52c-cb4f-467e-9044-1772b203fea8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.121363 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/b91d27f8-a52a-4a5a-a913-782a11c8e051-tmp-dir\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.121645 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f85237c3-327f-4a9e-8496-00c9087e5aa7-ca-trust-extracted\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.121833 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91d27f8-a52a-4a5a-a913-782a11c8e051-config\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.121939 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.621911726 +0000 UTC m=+102.915199118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.123713 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-certificates\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.125740 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f85237c3-327f-4a9e-8496-00c9087e5aa7-installation-pull-secrets\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.127887 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b91d27f8-a52a-4a5a-a913-782a11c8e051-serving-cert\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.128243 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-tls\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.134032 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"metrics-tls\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.157705 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.175014 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"ingress-operator-dockercfg-74nwh\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.194629 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221118 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221691 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6x7pt\" (UniqueName: \"kubernetes.io/projected/d9e554a8-6f94-4a46-acd7-3446e17421a5-kube-api-access-6x7pt\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221737 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-secret-volume\") pod \"collect-profiles-29567520-pnrmd\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221768 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-socket-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221788 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-tmp-dir\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221812 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-csi-data-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221832 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ed37e18-269e-4b50-a36a-0a723d5377d7-serving-cert\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221861 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6d5tr\" (UniqueName: \"kubernetes.io/projected/70789848-de09-49a1-8f00-e29b310194fa-kube-api-access-6d5tr\") pod \"multus-admission-controller-69db94689b-twxrl\" (UID: \"70789848-de09-49a1-8f00-e29b310194fa\") " pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221887 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/16a553dc-4b35-4c23-877a-4c7915ad94a4-tmpfs\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221909 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b9f2t\" (UniqueName: \"kubernetes.io/projected/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-kube-api-access-b9f2t\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221936 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pvhtm\" (UniqueName: \"kubernetes.io/projected/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-kube-api-access-pvhtm\") pod \"collect-profiles-29567520-pnrmd\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.221960 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ndrkw\" (UniqueName: \"kubernetes.io/projected/35de0468-bf61-4986-8473-77b7de39194a-kube-api-access-ndrkw\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222055 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-etcd-service-ca\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222089 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9e554a8-6f94-4a46-acd7-3446e17421a5-service-ca-bundle\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222115 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f9tc4\" (UniqueName: \"kubernetes.io/projected/16a553dc-4b35-4c23-877a-4c7915ad94a4-kube-api-access-f9tc4\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222141 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-config-volume\") pod \"collect-profiles-29567520-pnrmd\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222166 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/55cf9bde-83fe-4510-a1ae-acaf82c35b62-default-certificate\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222188 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/55cf9bde-83fe-4510-a1ae-acaf82c35b62-stats-auth\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222208 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d8ab52c-cb4f-467e-9044-1772b203fea8-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-x4xb5\" (UID: \"9d8ab52c-cb4f-467e-9044-1772b203fea8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222244 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bd48a05d-1bac-41bb-8b91-419cb4072895-profile-collector-cert\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222284 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9e554a8-6f94-4a46-acd7-3446e17421a5-config\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222310 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c5113992-e2f4-427c-acf3-55a4ef62376c-tmp-dir\") pod \"dns-operator-799b87ffcd-p2jqd\" (UID: \"c5113992-e2f4-427c-acf3-55a4ef62376c\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222334 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-etcd-client\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222375 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ed37e18-269e-4b50-a36a-0a723d5377d7-kube-api-access\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222398 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bntmk\" (UniqueName: \"kubernetes.io/projected/c5113992-e2f4-427c-acf3-55a4ef62376c-kube-api-access-bntmk\") pod \"dns-operator-799b87ffcd-p2jqd\" (UID: \"c5113992-e2f4-427c-acf3-55a4ef62376c\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222423 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cl6ml\" (UniqueName: \"kubernetes.io/projected/1a398b4e-beca-4bbc-a41a-76f738f128d8-kube-api-access-cl6ml\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222452 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8ab52c-cb4f-467e-9044-1772b203fea8-config\") pod \"kube-storage-version-migrator-operator-565b79b866-x4xb5\" (UID: \"9d8ab52c-cb4f-467e-9044-1772b203fea8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222477 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bd48a05d-1bac-41bb-8b91-419cb4072895-srv-cert\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222497 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-mountpoint-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222522 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/35de0468-bf61-4986-8473-77b7de39194a-tmp\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222551 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-etcd-ca\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222579 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-registration-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222603 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9e554a8-6f94-4a46-acd7-3446e17421a5-trusted-ca-bundle\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222628 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qsb42\" (UniqueName: \"kubernetes.io/projected/9462f609-9733-4abc-8f7c-f5fa0041f39d-kube-api-access-qsb42\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222652 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-prdxn\" (UniqueName: \"kubernetes.io/projected/55cf9bde-83fe-4510-a1ae-acaf82c35b62-kube-api-access-prdxn\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222678 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a398b4e-beca-4bbc-a41a-76f738f128d8-bound-sa-token\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222708 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ed37e18-269e-4b50-a36a-0a723d5377d7-config\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222755 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-plugins-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222783 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/58bdc39b-9471-4649-b1c4-45206dd8855d-tmp\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222809 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-94zxc\" (UniqueName: \"kubernetes.io/projected/1280d6d3-8430-41a6-9e76-5ddb21a0f44e-kube-api-access-94zxc\") pod \"machine-config-controller-f9cdd68f7-cpsp6\" (UID: \"1280d6d3-8430-41a6-9e76-5ddb21a0f44e\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222848 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5113992-e2f4-427c-acf3-55a4ef62376c-metrics-tls\") pod \"dns-operator-799b87ffcd-p2jqd\" (UID: \"c5113992-e2f4-427c-acf3-55a4ef62376c\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222873 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-serving-cert\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222899 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/7ed37e18-269e-4b50-a36a-0a723d5377d7-tmp-dir\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222921 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1280d6d3-8430-41a6-9e76-5ddb21a0f44e-mcc-auth-proxy-config\") pod \"machine-config-controller-f9cdd68f7-cpsp6\" (UID: \"1280d6d3-8430-41a6-9e76-5ddb21a0f44e\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222941 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55cf9bde-83fe-4510-a1ae-acaf82c35b62-service-ca-bundle\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222965 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a398b4e-beca-4bbc-a41a-76f738f128d8-trusted-ca\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.222988 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9e554a8-6f94-4a46-acd7-3446e17421a5-serving-cert\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223009 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-config\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223034 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bd48a05d-1bac-41bb-8b91-419cb4072895-tmpfs\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223058 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d4zg6\" (UniqueName: \"kubernetes.io/projected/9d8ab52c-cb4f-467e-9044-1772b203fea8-kube-api-access-d4zg6\") pod \"kube-storage-version-migrator-operator-565b79b866-x4xb5\" (UID: \"9d8ab52c-cb4f-467e-9044-1772b203fea8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223082 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a398b4e-beca-4bbc-a41a-76f738f128d8-metrics-tls\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223120 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/16a553dc-4b35-4c23-877a-4c7915ad94a4-apiservice-cert\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223143 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58bdc39b-9471-4649-b1c4-45206dd8855d-config\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223163 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/58bdc39b-9471-4649-b1c4-45206dd8855d-kube-api-access\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223189 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fz6bd\" (UniqueName: \"kubernetes.io/projected/bd48a05d-1bac-41bb-8b91-419cb4072895-kube-api-access-fz6bd\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223212 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35de0468-bf61-4986-8473-77b7de39194a-serving-cert\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223293 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/70789848-de09-49a1-8f00-e29b310194fa-webhook-certs\") pod \"multus-admission-controller-69db94689b-twxrl\" (UID: \"70789848-de09-49a1-8f00-e29b310194fa\") " pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223325 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55cf9bde-83fe-4510-a1ae-acaf82c35b62-metrics-certs\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223354 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1280d6d3-8430-41a6-9e76-5ddb21a0f44e-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-cpsp6\" (UID: \"1280d6d3-8430-41a6-9e76-5ddb21a0f44e\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223378 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/16a553dc-4b35-4c23-877a-4c7915ad94a4-webhook-cert\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223406 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35de0468-bf61-4986-8473-77b7de39194a-config\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223425 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58bdc39b-9471-4649-b1c4-45206dd8855d-serving-cert\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223494 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-mountpoint-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.223933 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bd48a05d-1bac-41bb-8b91-419cb4072895-tmpfs\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.224290 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c5113992-e2f4-427c-acf3-55a4ef62376c-tmp-dir\") pod \"dns-operator-799b87ffcd-p2jqd\" (UID: \"c5113992-e2f4-427c-acf3-55a4ef62376c\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.224478 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1280d6d3-8430-41a6-9e76-5ddb21a0f44e-mcc-auth-proxy-config\") pod \"machine-config-controller-f9cdd68f7-cpsp6\" (UID: \"1280d6d3-8430-41a6-9e76-5ddb21a0f44e\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.224743 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.224792 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/35de0468-bf61-4986-8473-77b7de39194a-tmp\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.224900 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9e554a8-6f94-4a46-acd7-3446e17421a5-config\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.225035 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-registration-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.225107 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ed37e18-269e-4b50-a36a-0a723d5377d7-config\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.225193 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-plugins-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.225764 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/58bdc39b-9471-4649-b1c4-45206dd8855d-tmp\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.226406 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9e554a8-6f94-4a46-acd7-3446e17421a5-trusted-ca-bundle\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.226611 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.227448 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/16a553dc-4b35-4c23-877a-4c7915ad94a4-tmpfs\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.227579 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/7ed37e18-269e-4b50-a36a-0a723d5377d7-tmp-dir\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.228970 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9e554a8-6f94-4a46-acd7-3446e17421a5-service-ca-bundle\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.230058 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35de0468-bf61-4986-8473-77b7de39194a-config\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.230805 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ed37e18-269e-4b50-a36a-0a723d5377d7-serving-cert\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.230924 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-csi-data-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.231023 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.731002938 +0000 UTC m=+103.024290110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.231290 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9462f609-9733-4abc-8f7c-f5fa0041f39d-socket-dir\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.236482 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5113992-e2f4-427c-acf3-55a4ef62376c-metrics-tls\") pod \"dns-operator-799b87ffcd-p2jqd\" (UID: \"c5113992-e2f4-427c-acf3-55a4ef62376c\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.237186 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-dockercfg-2wbn2\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.239986 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"trusted-ca\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.240628 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-tmp-dir\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.242092 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a398b4e-beca-4bbc-a41a-76f738f128d8-metrics-tls\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.242342 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9e554a8-6f94-4a46-acd7-3446e17421a5-serving-cert\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.244773 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35de0468-bf61-4986-8473-77b7de39194a-serving-cert\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.245958 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a398b4e-beca-4bbc-a41a-76f738f128d8-trusted-ca\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.255859 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.283297 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-config\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.290952 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58bdc39b-9471-4649-b1c4-45206dd8855d-config\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.293978 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-scheduler-operator-serving-cert\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.301613 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58bdc39b-9471-4649-b1c4-45206dd8855d-serving-cert\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.320897 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-controller-dockercfg-xnj77\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.329023 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c54d31e-9f4b-4071-8768-1eec361105d4-package-server-manager-serving-cert\") pod \"package-server-manager-77f986bd66-9zrjr\" (UID: \"7c54d31e-9f4b-4071-8768-1eec361105d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.329125 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.329184 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qnq6\" (UniqueName: \"kubernetes.io/projected/7c54d31e-9f4b-4071-8768-1eec361105d4-kube-api-access-4qnq6\") pod \"package-server-manager-77f986bd66-9zrjr\" (UID: \"7c54d31e-9f4b-4071-8768-1eec361105d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.329969 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.829951291 +0000 UTC m=+103.123238463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.335360 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mcc-proxy-tls\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.345310 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1280d6d3-8430-41a6-9e76-5ddb21a0f44e-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-cpsp6\" (UID: \"1280d6d3-8430-41a6-9e76-5ddb21a0f44e\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.356310 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-2h6bs\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.377617 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: W0321 00:13:09.386156 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbeedce19_bd88_4e81_a204_99cc121cd861.slice/crio-18fc8deffd324d2e3345470f1f3363b4fd9a51dcfc79b767820cb186dbbdf642 WatchSource:0}: Error finding container 18fc8deffd324d2e3345470f1f3363b4fd9a51dcfc79b767820cb186dbbdf642: Status 404 returned error can't find the container with id 18fc8deffd324d2e3345470f1f3363b4fd9a51dcfc79b767820cb186dbbdf642 Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.394511 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: W0321 00:13:09.415479 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfda377c1_1e51_4241_a939_a44cc536104d.slice/crio-0ed66714cc8752b8573a93027f04eb8fc96d7d1b075b6d6cb9be64caebdad364 WatchSource:0}: Error finding container 0ed66714cc8752b8573a93027f04eb8fc96d7d1b075b6d6cb9be64caebdad364: Status 404 returned error can't find the container with id 0ed66714cc8752b8573a93027f04eb8fc96d7d1b075b6d6cb9be64caebdad364 Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.415616 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.425038 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8ab52c-cb4f-467e-9044-1772b203fea8-config\") pod \"kube-storage-version-migrator-operator-565b79b866-x4xb5\" (UID: \"9d8ab52c-cb4f-467e-9044-1772b203fea8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.430739 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.430927 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.930895545 +0000 UTC m=+103.224182717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.431050 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.431132 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4qnq6\" (UniqueName: \"kubernetes.io/projected/7c54d31e-9f4b-4071-8768-1eec361105d4-kube-api-access-4qnq6\") pod \"package-server-manager-77f986bd66-9zrjr\" (UID: \"7c54d31e-9f4b-4071-8768-1eec361105d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.431370 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c54d31e-9f4b-4071-8768-1eec361105d4-package-server-manager-serving-cert\") pod \"package-server-manager-77f986bd66-9zrjr\" (UID: \"7c54d31e-9f4b-4071-8768-1eec361105d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.431781 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:09.931756021 +0000 UTC m=+103.225043203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.433504 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.447952 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d8ab52c-cb4f-467e-9044-1772b203fea8-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-x4xb5\" (UID: \"9d8ab52c-cb4f-467e-9044-1772b203fea8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.453555 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.473438 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-t8n29\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.493545 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"openshift-service-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.514629 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"kube-root-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.533273 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.533397 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.033366345 +0000 UTC m=+103.326653647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.534137 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-config\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.534672 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.535060 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.035050487 +0000 UTC m=+103.328337739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.545012 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-config-volume\") pod \"collect-profiles-29567520-pnrmd\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.554067 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"pprof-cert\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.560606 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-secret-volume\") pod \"collect-profiles-29567520-pnrmd\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.563083 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bd48a05d-1bac-41bb-8b91-419cb4072895-profile-collector-cert\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.574043 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.594417 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-kw8fx\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.614336 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-certs-default\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.630978 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/55cf9bde-83fe-4510-a1ae-acaf82c35b62-default-certificate\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.634848 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-stats-default\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.636561 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.636830 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.136786514 +0000 UTC m=+103.430073726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.637567 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.638926 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.138236498 +0000 UTC m=+103.431523680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.650157 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/55cf9bde-83fe-4510-a1ae-acaf82c35b62-stats-auth\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.655640 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-metrics-certs-default\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.664032 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55cf9bde-83fe-4510-a1ae-acaf82c35b62-metrics-certs\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.673052 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.693450 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"service-ca-bundle\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.696282 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55cf9bde-83fe-4510-a1ae-acaf82c35b62-service-ca-bundle\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.712574 5117 request.go:752] "Waited before sending request" delay="1.001871623s" reason="client-side throttling, not priority and fairness" verb="GET" URL="https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dcollect-profiles-dockercfg-vfqp6&limit=500&resourceVersion=0" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.716374 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-dockercfg-vfqp6\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.734494 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serviceaccount-dockercfg-4gqzj\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.738944 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.739394 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.239367818 +0000 UTC m=+103.532655000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.739999 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.740456 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.24044488 +0000 UTC m=+103.533732062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.754658 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"catalog-operator-serving-cert\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.758707 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.772498 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bd48a05d-1bac-41bb-8b91-419cb4072895-srv-cert\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.774453 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"openshift-service-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.794636 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"hostpath-provisioner\"/\"csi-hostpath-provisioner-sa-dockercfg-7dcws\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.813996 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"kube-root-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.834516 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"packageserver-service-cert\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.841522 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/16a553dc-4b35-4c23-877a-4c7915ad94a4-webhook-cert\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.841998 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/16a553dc-4b35-4c23-877a-4c7915ad94a4-apiservice-cert\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.842199 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.843620 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.343584601 +0000 UTC m=+103.636871783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.872288 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" event={"ID":"bed6c59c-20c9-41cd-92c8-6ca66de85fc8","Type":"ContainerStarted","Data":"f8f4acef8cb2dbadfef57969c6aa39ea20cb0fe456870928c4a73a7c812035d5"} Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.872403 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.872559 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.872704 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.872973 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.873150 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.879747 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ac-dockercfg-gj7jx\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.888707 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" event={"ID":"60516002-0c51-45a3-99a7-1ced5970206e","Type":"ContainerStarted","Data":"1b278493d55598e3ca0e2acf3cf536fcbaf10bf9c4aedc1421dd7115e96619d5"} Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.888778 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" event={"ID":"7d565d0a-934e-4b80-a621-3af94b7578a9","Type":"ContainerStarted","Data":"4c1fbcf7647d2908d04bf3fb6e8c66dd481abdbfb87e0c3cf43e516d1a0a694d"} Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.888809 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-74545575db-dmc74"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.894670 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-admission-controller-secret\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.902595 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/70789848-de09-49a1-8f00-e29b310194fa-webhook-certs\") pod \"multus-admission-controller-69db94689b-twxrl\" (UID: \"70789848-de09-49a1-8f00-e29b310194fa\") " pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.906634 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" event={"ID":"aed5803f-cd47-4e06-b5bf-059b2d51e70d","Type":"ContainerStarted","Data":"ffd31037dfd7d00a0316c9458de0c92f32bfd3986edb4d6b648bb2f252030c80"} Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.906690 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-m9nhf"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.907178 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.914855 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-ca-bundle\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.916473 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-etcd-ca\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.926069 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-67c89758df-fj8qh" event={"ID":"c68f3876-1fa5-43d6-b8a4-b278f689aaca","Type":"ContainerStarted","Data":"fa4208ecad7a9f9dcfa2e90091c2a7f84a7ce99236723cb5c36fb6787004dd8c"} Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.926284 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.926532 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-m9nhf" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.934641 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-dockercfg-4vdnc\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.936000 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mtrnq"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.936088 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-ll4kg"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.936419 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.945470 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dde4d157-ff00-4654-9b1e-f11894b4c047-auth-proxy-config\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.945884 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dde4d157-ff00-4654-9b1e-f11894b4c047-images\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.946055 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dde4d157-ff00-4654-9b1e-f11894b4c047-proxy-tls\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.946296 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rmsr\" (UniqueName: \"kubernetes.io/projected/dde4d157-ff00-4654-9b1e-f11894b4c047-kube-api-access-8rmsr\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.946468 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:09 crc kubenswrapper[5117]: E0321 00:13:09.947102 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.447076352 +0000 UTC m=+103.740363704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.950829 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-67c89758df-fj8qh" event={"ID":"c68f3876-1fa5-43d6-b8a4-b278f689aaca","Type":"ContainerStarted","Data":"87219001c2256af3b36305fe7fe9bfa0068c7ee968654ea6324e3522d7049474"} Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.950888 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.951088 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.953790 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-serving-cert\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.970842 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-serving-cert\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.972467 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.972599 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.974512 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992664 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992703 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" event={"ID":"828ac2ae-7b43-4714-9cb9-a86252d5deeb","Type":"ContainerStarted","Data":"5363f9fb6b2433e80f967a3f23e41e183a646fe86e0fbe60792e5b0da35c69ca"} Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992724 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" event={"ID":"a5a8eb05-2bff-483b-b42f-03b818f9a058","Type":"ContainerStarted","Data":"df46ff372feb0a56a5e8d86f9ba8582bbe49ad744382c4df90099fdf1892bda8"} Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992740 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-799b87ffcd-p2jqd"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992752 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29567520-5stkm"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992763 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992775 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-755bb95488-d9bzl"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992784 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-gjth4"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992794 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992808 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992818 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992833 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992842 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992871 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-9ddfb9f55-nsfg8"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992881 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64d44f6ddf-dg2qw"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992891 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.992902 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-gbnbk"] Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.993947 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" Mar 21 00:13:09 crc kubenswrapper[5117]: I0321 00:13:09.994268 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-service-ca-bundle\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.000956 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-etcd-service-ca\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.014066 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-client\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.019635 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2n4q9"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.019692 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-74545575db-dmc74"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.019708 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.019723 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.019739 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-jjwjr"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.020590 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.026216 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-etcd-client\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.028724 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.028754 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.028767 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.028777 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-m9nhf"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.028787 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.028803 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-6qjwh"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.029153 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.034422 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039005 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-69db94689b-twxrl"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039034 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039050 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-ll4kg"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039063 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039074 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039084 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039095 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039106 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gbnbk"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039192 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-755bb95488-d9bzl"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039204 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-dx9lp"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039214 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-67c89758df-fj8qh"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039225 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039224 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039238 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-9ddfb9f55-nsfg8"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039264 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039273 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mtrnq"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039283 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039293 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29567520-5stkm"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039303 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039312 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-5777786469-xqfxc"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039321 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-747b44746d-7zctx"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039331 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64d44f6ddf-dg2qw"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.039344 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.047940 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.048221 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.548189501 +0000 UTC m=+103.841476673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.049444 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.049679 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.049746 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/4db781b1-fcbc-44e9-aa0f-56184c19f397-tmp-dir\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.049774 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-certs\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.049816 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ee00846f-ea12-4777-a333-942c9e5e7cea-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-75ffdb6fcd-qt6hv\" (UID: \"ee00846f-ea12-4777-a333-942c9e5e7cea\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.049880 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79xmn\" (UniqueName: \"kubernetes.io/projected/384266b7-4505-4bb0-9330-ad1a2eebba3c-kube-api-access-79xmn\") pod \"service-ca-operator-5b9c976747-n7fn2\" (UID: \"384266b7-4505-4bb0-9330-ad1a2eebba3c\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.049931 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5m9d\" (UniqueName: \"kubernetes.io/projected/7b571f34-0996-43b7-b450-cb7273d9ab00-kube-api-access-p5m9d\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.049954 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-ready\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.049977 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ltk9\" (UniqueName: \"kubernetes.io/projected/724bd3f7-7190-428f-8ab9-05416c2c1096-kube-api-access-7ltk9\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050005 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/af2a44fd-39a9-41fc-9d1f-640e828d6a16-tmp\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050031 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qwbc\" (UniqueName: \"kubernetes.io/projected/4db781b1-fcbc-44e9-aa0f-56184c19f397-kube-api-access-8qwbc\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050061 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/724bd3f7-7190-428f-8ab9-05416c2c1096-tmpfs\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050084 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050104 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g8l7\" (UniqueName: \"kubernetes.io/projected/af2a44fd-39a9-41fc-9d1f-640e828d6a16-kube-api-access-9g8l7\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050170 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4db781b1-fcbc-44e9-aa0f-56184c19f397-config-volume\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050512 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tt25\" (UniqueName: \"kubernetes.io/projected/ee00846f-ea12-4777-a333-942c9e5e7cea-kube-api-access-8tt25\") pod \"control-plane-machine-set-operator-75ffdb6fcd-qt6hv\" (UID: \"ee00846f-ea12-4777-a333-942c9e5e7cea\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050600 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/258bf46a-4978-43d1-8579-46bfe65d8ba8-signing-key\") pod \"service-ca-74545575db-dmc74\" (UID: \"258bf46a-4978-43d1-8579-46bfe65d8ba8\") " pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050626 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/724bd3f7-7190-428f-8ab9-05416c2c1096-srv-cert\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050672 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/384266b7-4505-4bb0-9330-ad1a2eebba3c-config\") pod \"service-ca-operator-5b9c976747-n7fn2\" (UID: \"384266b7-4505-4bb0-9330-ad1a2eebba3c\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050698 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78edefc3-3169-4348-b76a-c255508c077f-cert\") pod \"ingress-canary-m9nhf\" (UID: \"78edefc3-3169-4348-b76a-c255508c077f\") " pod="openshift-ingress-canary/ingress-canary-m9nhf" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050744 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050773 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050803 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050826 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dde4d157-ff00-4654-9b1e-f11894b4c047-auth-proxy-config\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050869 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bww62\" (UniqueName: \"kubernetes.io/projected/78edefc3-3169-4348-b76a-c255508c077f-kube-api-access-bww62\") pod \"ingress-canary-m9nhf\" (UID: \"78edefc3-3169-4348-b76a-c255508c077f\") " pod="openshift-ingress-canary/ingress-canary-m9nhf" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050893 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dde4d157-ff00-4654-9b1e-f11894b4c047-images\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050918 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/384266b7-4505-4bb0-9330-ad1a2eebba3c-serving-cert\") pod \"service-ca-operator-5b9c976747-n7fn2\" (UID: \"384266b7-4505-4bb0-9330-ad1a2eebba3c\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.050989 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbmxh\" (UniqueName: \"kubernetes.io/projected/b77f02bf-fa45-4ac2-bf1c-aa01afc3f637-kube-api-access-dbmxh\") pod \"migrator-866fcbc849-7wxvx\" (UID: \"b77f02bf-fa45-4ac2-bf1c-aa01afc3f637\") " pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051052 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6hmq\" (UniqueName: \"kubernetes.io/projected/258bf46a-4978-43d1-8579-46bfe65d8ba8-kube-api-access-h6hmq\") pod \"service-ca-74545575db-dmc74\" (UID: \"258bf46a-4978-43d1-8579-46bfe65d8ba8\") " pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051078 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dchfn\" (UniqueName: \"kubernetes.io/projected/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-kube-api-access-dchfn\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051113 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051136 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4db781b1-fcbc-44e9-aa0f-56184c19f397-metrics-tls\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051164 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dde4d157-ff00-4654-9b1e-f11894b4c047-proxy-tls\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051457 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051532 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-node-bootstrap-token\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051637 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/258bf46a-4978-43d1-8579-46bfe65d8ba8-signing-cabundle\") pod \"service-ca-74545575db-dmc74\" (UID: \"258bf46a-4978-43d1-8579-46bfe65d8ba8\") " pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051688 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8rmsr\" (UniqueName: \"kubernetes.io/projected/dde4d157-ff00-4654-9b1e-f11894b4c047-kube-api-access-8rmsr\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051723 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/724bd3f7-7190-428f-8ab9-05416c2c1096-profile-collector-cert\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051792 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.051816 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dde4d157-ff00-4654-9b1e-f11894b4c047-auth-proxy-config\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.052361 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.552338218 +0000 UTC m=+103.845625410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.053660 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-config\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.054376 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-config\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.092101 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-bound-sa-token\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.108976 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2677t\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-kube-api-access-2677t\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.129914 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b91d27f8-a52a-4a5a-a913-782a11c8e051-kube-api-access\") pod \"kube-apiserver-operator-575994946d-lkslg\" (UID: \"b91d27f8-a52a-4a5a-a913-782a11c8e051\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.148799 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9tc4\" (UniqueName: \"kubernetes.io/projected/16a553dc-4b35-4c23-877a-4c7915ad94a4-kube-api-access-f9tc4\") pod \"packageserver-7d4fc7d867-k4rmn\" (UID: \"16a553dc-4b35-4c23-877a-4c7915ad94a4\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.149288 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" event={"ID":"a5a8eb05-2bff-483b-b42f-03b818f9a058","Type":"ContainerStarted","Data":"55b428f845a56f759367d07dee79ae047819a764952b0ed1e3738688f5053052"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.149563 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.152294 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" event={"ID":"828ac2ae-7b43-4714-9cb9-a86252d5deeb","Type":"ContainerStarted","Data":"eae4734986c4bb994dca2056feb8ad2f65a1a7472072fd2498e0692a80376f5a"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.152370 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.152364 5117 patch_prober.go:28] interesting pod/controller-manager-65b6cccf98-dx9lp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.152490 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.652465826 +0000 UTC m=+103.945752998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.152510 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" podUID="a5a8eb05-2bff-483b-b42f-03b818f9a058" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.152984 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/384266b7-4505-4bb0-9330-ad1a2eebba3c-config\") pod \"service-ca-operator-5b9c976747-n7fn2\" (UID: \"384266b7-4505-4bb0-9330-ad1a2eebba3c\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.153024 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78edefc3-3169-4348-b76a-c255508c077f-cert\") pod \"ingress-canary-m9nhf\" (UID: \"78edefc3-3169-4348-b76a-c255508c077f\") " pod="openshift-ingress-canary/ingress-canary-m9nhf" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.153154 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.153280 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bww62\" (UniqueName: \"kubernetes.io/projected/78edefc3-3169-4348-b76a-c255508c077f-kube-api-access-bww62\") pod \"ingress-canary-m9nhf\" (UID: \"78edefc3-3169-4348-b76a-c255508c077f\") " pod="openshift-ingress-canary/ingress-canary-m9nhf" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.153309 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/384266b7-4505-4bb0-9330-ad1a2eebba3c-serving-cert\") pod \"service-ca-operator-5b9c976747-n7fn2\" (UID: \"384266b7-4505-4bb0-9330-ad1a2eebba3c\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.153343 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dbmxh\" (UniqueName: \"kubernetes.io/projected/b77f02bf-fa45-4ac2-bf1c-aa01afc3f637-kube-api-access-dbmxh\") pod \"migrator-866fcbc849-7wxvx\" (UID: \"b77f02bf-fa45-4ac2-bf1c-aa01afc3f637\") " pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.153691 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" event={"ID":"9f6e0dec-9f57-403b-893c-5a30c576a799","Type":"ContainerStarted","Data":"268849b5d203095fee07bbc60f91a29e7fade07256f6787d5b89936fe038b834"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.155436 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.155777 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-h6hmq\" (UniqueName: \"kubernetes.io/projected/258bf46a-4978-43d1-8579-46bfe65d8ba8-kube-api-access-h6hmq\") pod \"service-ca-74545575db-dmc74\" (UID: \"258bf46a-4978-43d1-8579-46bfe65d8ba8\") " pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.155878 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dchfn\" (UniqueName: \"kubernetes.io/projected/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-kube-api-access-dchfn\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.155951 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4db781b1-fcbc-44e9-aa0f-56184c19f397-metrics-tls\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.155990 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156030 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-node-bootstrap-token\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156075 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/258bf46a-4978-43d1-8579-46bfe65d8ba8-signing-cabundle\") pod \"service-ca-74545575db-dmc74\" (UID: \"258bf46a-4978-43d1-8579-46bfe65d8ba8\") " pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156122 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/724bd3f7-7190-428f-8ab9-05416c2c1096-profile-collector-cert\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156190 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156299 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156347 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/4db781b1-fcbc-44e9-aa0f-56184c19f397-tmp-dir\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156383 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-certs\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156411 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ee00846f-ea12-4777-a333-942c9e5e7cea-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-75ffdb6fcd-qt6hv\" (UID: \"ee00846f-ea12-4777-a333-942c9e5e7cea\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156459 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79xmn\" (UniqueName: \"kubernetes.io/projected/384266b7-4505-4bb0-9330-ad1a2eebba3c-kube-api-access-79xmn\") pod \"service-ca-operator-5b9c976747-n7fn2\" (UID: \"384266b7-4505-4bb0-9330-ad1a2eebba3c\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.156497 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p5m9d\" (UniqueName: \"kubernetes.io/projected/7b571f34-0996-43b7-b450-cb7273d9ab00-kube-api-access-p5m9d\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.159967 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.160547 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/4db781b1-fcbc-44e9-aa0f-56184c19f397-tmp-dir\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.160935 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.660914254 +0000 UTC m=+103.954201606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161135 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-ready\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161192 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7ltk9\" (UniqueName: \"kubernetes.io/projected/724bd3f7-7190-428f-8ab9-05416c2c1096-kube-api-access-7ltk9\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161237 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/af2a44fd-39a9-41fc-9d1f-640e828d6a16-tmp\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161296 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8qwbc\" (UniqueName: \"kubernetes.io/projected/4db781b1-fcbc-44e9-aa0f-56184c19f397-kube-api-access-8qwbc\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161339 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/724bd3f7-7190-428f-8ab9-05416c2c1096-tmpfs\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161371 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161398 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9g8l7\" (UniqueName: \"kubernetes.io/projected/af2a44fd-39a9-41fc-9d1f-640e828d6a16-kube-api-access-9g8l7\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161424 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4db781b1-fcbc-44e9-aa0f-56184c19f397-config-volume\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161452 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8tt25\" (UniqueName: \"kubernetes.io/projected/ee00846f-ea12-4777-a333-942c9e5e7cea-kube-api-access-8tt25\") pod \"control-plane-machine-set-operator-75ffdb6fcd-qt6hv\" (UID: \"ee00846f-ea12-4777-a333-942c9e5e7cea\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161496 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/258bf46a-4978-43d1-8579-46bfe65d8ba8-signing-key\") pod \"service-ca-74545575db-dmc74\" (UID: \"258bf46a-4978-43d1-8579-46bfe65d8ba8\") " pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161526 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/724bd3f7-7190-428f-8ab9-05416c2c1096-srv-cert\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.161781 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/482a0770-6dc5-4682-bf1e-e3c085129f33-metrics-certs\") pod \"network-metrics-daemon-qp2t9\" (UID: \"482a0770-6dc5-4682-bf1e-e3c085129f33\") " pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.162079 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/724bd3f7-7190-428f-8ab9-05416c2c1096-tmpfs\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.162380 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-ready\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.162576 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/af2a44fd-39a9-41fc-9d1f-640e828d6a16-tmp\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.164779 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" event={"ID":"3a9dda43-9285-4470-8175-f919dd943c17","Type":"ContainerStarted","Data":"0848ebfe9632e31d54037374b28c7c4f626b140dcce746569fb03114d02d2216"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.166577 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/724bd3f7-7190-428f-8ab9-05416c2c1096-profile-collector-cert\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.168888 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ed37e18-269e-4b50-a36a-0a723d5377d7-kube-api-access\") pod \"kube-controller-manager-operator-69d5f845f8-nj24k\" (UID: \"7ed37e18-269e-4b50-a36a-0a723d5377d7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.172857 5117 generic.go:358] "Generic (PLEG): container finished" podID="7d565d0a-934e-4b80-a621-3af94b7578a9" containerID="3b2ac630f0a110426280b46452926aef4128b54e6523314b13ed1dcd6582a309" exitCode=0 Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.172965 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" event={"ID":"7d565d0a-934e-4b80-a621-3af94b7578a9","Type":"ContainerDied","Data":"3b2ac630f0a110426280b46452926aef4128b54e6523314b13ed1dcd6582a309"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.180293 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" event={"ID":"a29e146a-be7b-49f8-b97b-31f4c976afb5","Type":"ContainerStarted","Data":"34c3f5ed25dce25348c1c06732daffb81ae4d5ab489fe4b58efdeaf4a4996683"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.183971 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" event={"ID":"beedce19-bd88-4e81-a204-99cc121cd861","Type":"ContainerStarted","Data":"18fc8deffd324d2e3345470f1f3363b4fd9a51dcfc79b767820cb186dbbdf642"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.185786 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" event={"ID":"aed5803f-cd47-4e06-b5bf-059b2d51e70d","Type":"ContainerStarted","Data":"375994de4f379c09e8d40c09c060a02a4de9d51f5c5b43aa57b884658f0c052e"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.187033 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-747b44746d-7zctx" event={"ID":"6c43692a-55c8-4d6a-bd36-9c3268722421","Type":"ContainerStarted","Data":"634d4c9c5a5c06d023ff64f88c4c76ab08f7a78746e85c266e9ed2eecb83ddb0"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.188179 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29567520-5stkm" event={"ID":"79fcf4ae-c05f-492c-88f0-ec441eb255aa","Type":"ContainerStarted","Data":"f878db1c1959de3cbdb4308c50e29846a88745aecb3b9b05353134cbf045d89f"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.189350 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" event={"ID":"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2","Type":"ContainerStarted","Data":"2db8c230993a06b01a23be9a47ff1dfdea780e559cef94464edba9e4a2f70d99"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.193465 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bntmk\" (UniqueName: \"kubernetes.io/projected/c5113992-e2f4-427c-acf3-55a4ef62376c-kube-api-access-bntmk\") pod \"dns-operator-799b87ffcd-p2jqd\" (UID: \"c5113992-e2f4-427c-acf3-55a4ef62376c\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.194601 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" event={"ID":"bed6c59c-20c9-41cd-92c8-6ca66de85fc8","Type":"ContainerStarted","Data":"09561e725fb8c46d607dad059ebbc8ecf5fe7957b9f65d8213a8f4f07abb3810"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.197575 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64d44f6ddf-dg2qw" event={"ID":"fda377c1-1e51-4241-a939-a44cc536104d","Type":"ContainerStarted","Data":"0ed66714cc8752b8573a93027f04eb8fc96d7d1b075b6d6cb9be64caebdad364"} Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.205430 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.207426 5117 patch_prober.go:28] interesting pod/console-operator-67c89758df-fj8qh container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.207490 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-67c89758df-fj8qh" podUID="c68f3876-1fa5-43d6-b8a4-b278f689aaca" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.210477 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl6ml\" (UniqueName: \"kubernetes.io/projected/1a398b4e-beca-4bbc-a41a-76f738f128d8-kube-api-access-cl6ml\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.232202 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-94zxc\" (UniqueName: \"kubernetes.io/projected/1280d6d3-8430-41a6-9e76-5ddb21a0f44e-kube-api-access-94zxc\") pod \"machine-config-controller-f9cdd68f7-cpsp6\" (UID: \"1280d6d3-8430-41a6-9e76-5ddb21a0f44e\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.250045 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.252812 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9f2t\" (UniqueName: \"kubernetes.io/projected/e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3-kube-api-access-b9f2t\") pod \"etcd-operator-69b85846b6-cvk4t\" (UID: \"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.272994 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.273410 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.77337804 +0000 UTC m=+104.066665202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.273787 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.276517 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.776504865 +0000 UTC m=+104.069792037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.277367 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/58bdc39b-9471-4649-b1c4-45206dd8855d-kube-api-access\") pod \"openshift-kube-scheduler-operator-54f497555d-4xkzx\" (UID: \"58bdc39b-9471-4649-b1c4-45206dd8855d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.285782 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.294136 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsb42\" (UniqueName: \"kubernetes.io/projected/9462f609-9733-4abc-8f7c-f5fa0041f39d-kube-api-access-qsb42\") pod \"csi-hostpathplugin-2n4q9\" (UID: \"9462f609-9733-4abc-8f7c-f5fa0041f39d\") " pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.313663 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qp2t9" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.320058 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-prdxn\" (UniqueName: \"kubernetes.io/projected/55cf9bde-83fe-4510-a1ae-acaf82c35b62-kube-api-access-prdxn\") pod \"router-default-68cf44c8b8-zkpbt\" (UID: \"55cf9bde-83fe-4510-a1ae-acaf82c35b62\") " pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.336077 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz6bd\" (UniqueName: \"kubernetes.io/projected/bd48a05d-1bac-41bb-8b91-419cb4072895-kube-api-access-fz6bd\") pod \"catalog-operator-75ff9f647d-hbjn2\" (UID: \"bd48a05d-1bac-41bb-8b91-419cb4072895\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.351760 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.354622 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.355520 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvhtm\" (UniqueName: \"kubernetes.io/projected/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-kube-api-access-pvhtm\") pod \"collect-profiles-29567520-pnrmd\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.363233 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.375656 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.375888 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndrkw\" (UniqueName: \"kubernetes.io/projected/35de0468-bf61-4986-8473-77b7de39194a-kube-api-access-ndrkw\") pod \"openshift-controller-manager-operator-686468bdd5-wmwn8\" (UID: \"35de0468-bf61-4986-8473-77b7de39194a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.377611 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"package-server-manager-serving-cert\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.377691 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.378102 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.878080338 +0000 UTC m=+104.171367510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.392538 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c54d31e-9f4b-4071-8768-1eec361105d4-package-server-manager-serving-cert\") pod \"package-server-manager-77f986bd66-9zrjr\" (UID: \"7c54d31e-9f4b-4071-8768-1eec361105d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.395652 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.443162 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a398b4e-beca-4bbc-a41a-76f738f128d8-bound-sa-token\") pod \"ingress-operator-6b9cb4dbcf-zgbp6\" (UID: \"1a398b4e-beca-4bbc-a41a-76f738f128d8\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.445284 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.452147 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d5tr\" (UniqueName: \"kubernetes.io/projected/70789848-de09-49a1-8f00-e29b310194fa-kube-api-access-6d5tr\") pod \"multus-admission-controller-69db94689b-twxrl\" (UID: \"70789848-de09-49a1-8f00-e29b310194fa\") " pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.457078 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4zg6\" (UniqueName: \"kubernetes.io/projected/9d8ab52c-cb4f-467e-9044-1772b203fea8-kube-api-access-d4zg6\") pod \"kube-storage-version-migrator-operator-565b79b866-x4xb5\" (UID: \"9d8ab52c-cb4f-467e-9044-1772b203fea8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.458618 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.466169 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.468967 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x7pt\" (UniqueName: \"kubernetes.io/projected/d9e554a8-6f94-4a46-acd7-3446e17421a5-kube-api-access-6x7pt\") pod \"authentication-operator-7f5c659b84-cn4pf\" (UID: \"d9e554a8-6f94-4a46-acd7-3446e17421a5\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.479883 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.480348 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:10.980327841 +0000 UTC m=+104.273615013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.482089 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.494602 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qnq6\" (UniqueName: \"kubernetes.io/projected/7c54d31e-9f4b-4071-8768-1eec361105d4-kube-api-access-4qnq6\") pod \"package-server-manager-77f986bd66-9zrjr\" (UID: \"7c54d31e-9f4b-4071-8768-1eec361105d4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.515988 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.538643 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.539923 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mco-proxy-tls\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.540346 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.566400 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-images\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.566404 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dde4d157-ff00-4654-9b1e-f11894b4c047-proxy-tls\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.571285 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.573494 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dde4d157-ff00-4654-9b1e-f11894b4c047-images\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.574494 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-dockercfg-sw6nc\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.582988 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.585862 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.586274 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.086237487 +0000 UTC m=+104.379524659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.595038 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.614412 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.617456 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-console\"/\"networking-console-plugin\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.620600 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.623881 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.632460 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.640244 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"networking-console-plugin-cert\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.661662 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.667390 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.674135 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-dockercfg-bjqfd\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.687477 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.688123 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.188101449 +0000 UTC m=+104.481388621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.693641 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-qp2t9"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.695966 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"serving-cert\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.700844 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-799b87ffcd-p2jqd"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.714437 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-config\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.724231 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/384266b7-4505-4bb0-9330-ad1a2eebba3c-config\") pod \"service-ca-operator-5b9c976747-n7fn2\" (UID: \"384266b7-4505-4bb0-9330-ad1a2eebba3c\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.734882 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"kube-root-ca.crt\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.737618 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/384266b7-4505-4bb0-9330-ad1a2eebba3c-serving-cert\") pod \"service-ca-operator-5b9c976747-n7fn2\" (UID: \"384266b7-4505-4bb0-9330-ad1a2eebba3c\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.747696 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t"] Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.751639 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.753329 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.773463 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-bgxvm\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.791174 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.791376 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.291342572 +0000 UTC m=+104.584629754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.792008 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.792600 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.292582939 +0000 UTC m=+104.585870111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.794141 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.803932 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.814568 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.818175 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/258bf46a-4978-43d1-8579-46bfe65d8ba8-signing-cabundle\") pod \"service-ca-74545575db-dmc74\" (UID: \"258bf46a-4978-43d1-8579-46bfe65d8ba8\") " pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.820803 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.833732 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.848974 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/258bf46a-4978-43d1-8579-46bfe65d8ba8-signing-key\") pod \"service-ca-74545575db-dmc74\" (UID: \"258bf46a-4978-43d1-8579-46bfe65d8ba8\") " pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.854079 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.875553 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.893420 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.893592 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.393562664 +0000 UTC m=+104.686849836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.894158 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.894536 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.394529104 +0000 UTC m=+104.687816276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.894699 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.921241 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9pgs7\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.934163 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serving-cert\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.946047 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/724bd3f7-7190-428f-8ab9-05416c2c1096-srv-cert\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.951407 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78edefc3-3169-4348-b76a-c255508c077f-cert\") pod \"ingress-canary-m9nhf\" (UID: \"78edefc3-3169-4348-b76a-c255508c077f\") " pod="openshift-ingress-canary/ingress-canary-m9nhf" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.951916 5117 request.go:752] "Waited before sending request" delay="1.000114541s" reason="client-side throttling, not priority and fairness" verb="GET" URL="https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/secrets?fieldSelector=metadata.name%3Dmarketplace-operator-dockercfg-2cfkp&limit=500&resourceVersion=0" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.954121 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-dockercfg-2cfkp\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.963963 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.977526 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-metrics\"" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.994071 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.996174 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:10 crc kubenswrapper[5117]: E0321 00:13:10.996489 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.496472288 +0000 UTC m=+104.789759460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:10 crc kubenswrapper[5117]: I0321 00:13:10.996645 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.023627 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"marketplace-trusted-ca\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.033011 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.036018 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.056094 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.081696 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-kknhg\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.095350 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.098481 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.100633 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.600614519 +0000 UTC m=+104.893901691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.121336 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-dockercfg-gnx66\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.136996 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-tls\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.151018 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ee00846f-ea12-4777-a333-942c9e5e7cea-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-75ffdb6fcd-qt6hv\" (UID: \"ee00846f-ea12-4777-a333-942c9e5e7cea\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.153722 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.159886 5117 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.159983 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4db781b1-fcbc-44e9-aa0f-56184c19f397-metrics-tls podName:4db781b1-fcbc-44e9-aa0f-56184c19f397 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.659956412 +0000 UTC m=+104.953243584 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/4db781b1-fcbc-44e9-aa0f-56184c19f397-metrics-tls") pod "dns-default-gbnbk" (UID: "4db781b1-fcbc-44e9-aa0f-56184c19f397") : failed to sync secret cache: timed out waiting for the condition Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.160084 5117 secret.go:189] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.160202 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-node-bootstrap-token podName:7b571f34-0996-43b7-b450-cb7273d9ab00 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.660180269 +0000 UTC m=+104.953467441 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-node-bootstrap-token") pod "machine-config-server-6qjwh" (UID: "7b571f34-0996-43b7-b450-cb7273d9ab00") : failed to sync secret cache: timed out waiting for the condition Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.162476 5117 secret.go:189] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.162639 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-certs podName:7b571f34-0996-43b7-b450-cb7273d9ab00 nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.662622553 +0000 UTC m=+104.955909725 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-certs") pod "machine-config-server-6qjwh" (UID: "7b571f34-0996-43b7-b450-cb7273d9ab00") : failed to sync secret cache: timed out waiting for the condition Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.162643 5117 configmap.go:193] Couldn't get configMap openshift-multus/cni-sysctl-allowlist: failed to sync configmap cache: timed out waiting for the condition Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.162702 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-cni-sysctl-allowlist podName:ee9b5fe1-7a59-49c5-b486-a6d83d56728c nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.662693195 +0000 UTC m=+104.955980367 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-sysctl-allowlist" (UniqueName: "kubernetes.io/configmap/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-cni-sysctl-allowlist") pod "cni-sysctl-allowlist-ds-jjwjr" (UID: "ee9b5fe1-7a59-49c5-b486-a6d83d56728c") : failed to sync configmap cache: timed out waiting for the condition Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.162960 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4db781b1-fcbc-44e9-aa0f-56184c19f397-config-volume\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.175261 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.196862 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-kpvmz\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.200668 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.200799 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.700778469 +0000 UTC m=+104.994065641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.201161 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.201667 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.701642675 +0000 UTC m=+104.994929847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.221678 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-sysctl-allowlist\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.224772 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" event={"ID":"aed5803f-cd47-4e06-b5bf-059b2d51e70d","Type":"ContainerStarted","Data":"7c86f31cb22c24d379692cf29b95e688a6ee0a18ac0fcb9e012bcbe8a3781d2d"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.232654 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" event={"ID":"c5113992-e2f4-427c-acf3-55a4ef62376c","Type":"ContainerStarted","Data":"ce1b1e3b853b9067ddc6cd297705e77bef170cca4342df3d20fb0510d7376098"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.233562 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-dockercfg-dzw6b\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.234800 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-747b44746d-7zctx" event={"ID":"6c43692a-55c8-4d6a-bd36-9c3268722421","Type":"ContainerStarted","Data":"8d96f4012bf6304db8a60a8b2325dc83430d65e2933cd34f1e25998ba7d63d6e"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.235790 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/downloads-747b44746d-7zctx" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.237524 5117 patch_prober.go:28] interesting pod/downloads-747b44746d-7zctx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.237592 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-747b44746d-7zctx" podUID="6c43692a-55c8-4d6a-bd36-9c3268722421" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.241014 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29567520-5stkm" event={"ID":"79fcf4ae-c05f-492c-88f0-ec441eb255aa","Type":"ContainerStarted","Data":"ed9b114390e516515e6dce7ca532581b221ff2031f6a039a59d36381df3949f4"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.256527 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"node-bootstrapper-token\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.268895 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" event={"ID":"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2","Type":"ContainerStarted","Data":"4895ca8aa86f5389620bf2a7362ecb3f7ab72511f6c03a9de0137fccce21b86a"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.270285 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.271578 5117 patch_prober.go:28] interesting pod/oauth-openshift-66458b6674-mtrnq container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.271627 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" podUID="7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.281017 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-tls\"" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.286632 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64d44f6ddf-dg2qw" event={"ID":"fda377c1-1e51-4241-a939-a44cc536104d","Type":"ContainerStarted","Data":"0b7905fe95b70c856b16492e2b10eb280840b02b6f9947b3490cec5f30bcdb76"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.305916 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.307176 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.807142338 +0000 UTC m=+105.100429510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.318586 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rmsr\" (UniqueName: \"kubernetes.io/projected/dde4d157-ff00-4654-9b1e-f11894b4c047-kube-api-access-8rmsr\") pod \"machine-config-operator-67c9d58cbb-l5tfb\" (UID: \"dde4d157-ff00-4654-9b1e-f11894b4c047\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.324624 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" event={"ID":"55cf9bde-83fe-4510-a1ae-acaf82c35b62","Type":"ContainerStarted","Data":"abed5845fbcbd42aa68c6e5090e6a71b5b517b2363b5316d5a4a37c37b1942ed"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.324695 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" event={"ID":"55cf9bde-83fe-4510-a1ae-acaf82c35b62","Type":"ContainerStarted","Data":"f2f21900e0cf94840bef18830718e61f6c30257c8d9b977cb9d625837568358b"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.335164 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" event={"ID":"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3","Type":"ContainerStarted","Data":"a4e8486a6b42a254cb0ab1e815a4b5080cacdb1b5e8a6a8b30a13157b1d36a4e"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.337661 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" event={"ID":"828ac2ae-7b43-4714-9cb9-a86252d5deeb","Type":"ContainerStarted","Data":"1daa5fdfc079ddb1fa5d00aa9d86a8503d835fd3a8b1d485f008bdf644c90573"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.340884 5117 generic.go:358] "Generic (PLEG): container finished" podID="9f6e0dec-9f57-403b-893c-5a30c576a799" containerID="a3e6b70affb5a82d6865adf6e2a0883670d9e5b59560efe98a015595c0bd9583" exitCode=0 Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.340989 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" event={"ID":"9f6e0dec-9f57-403b-893c-5a30c576a799","Type":"ContainerDied","Data":"a3e6b70affb5a82d6865adf6e2a0883670d9e5b59560efe98a015595c0bd9583"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.356373 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bww62\" (UniqueName: \"kubernetes.io/projected/78edefc3-3169-4348-b76a-c255508c077f-kube-api-access-bww62\") pod \"ingress-canary-m9nhf\" (UID: \"78edefc3-3169-4348-b76a-c255508c077f\") " pod="openshift-ingress-canary/ingress-canary-m9nhf" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.358572 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.367361 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbmxh\" (UniqueName: \"kubernetes.io/projected/b77f02bf-fa45-4ac2-bf1c-aa01afc3f637-kube-api-access-dbmxh\") pod \"migrator-866fcbc849-7wxvx\" (UID: \"b77f02bf-fa45-4ac2-bf1c-aa01afc3f637\") " pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.381880 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" event={"ID":"3a9dda43-9285-4470-8175-f919dd943c17","Type":"ContainerStarted","Data":"9d438f9a13529b7ac453e9a88b309fcf9eb704035da9a9c77a96a9936229eedd"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.396648 5117 generic.go:358] "Generic (PLEG): container finished" podID="60516002-0c51-45a3-99a7-1ced5970206e" containerID="14f5ce1ebea792fee9d1ca75be2e1e9bc4fadfe1e9cb5d265c26661e3d671a1c" exitCode=0 Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.396977 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" event={"ID":"60516002-0c51-45a3-99a7-1ced5970206e","Type":"ContainerDied","Data":"14f5ce1ebea792fee9d1ca75be2e1e9bc4fadfe1e9cb5d265c26661e3d671a1c"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.407099 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5m9d\" (UniqueName: \"kubernetes.io/projected/7b571f34-0996-43b7-b450-cb7273d9ab00-kube-api-access-p5m9d\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.408377 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" event={"ID":"7d565d0a-934e-4b80-a621-3af94b7578a9","Type":"ContainerStarted","Data":"b171e6e773516c0b8c4dc4f65059515dc0580da5e1c0da4fe925f296e895f515"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.411297 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.413301 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:11.913285151 +0000 UTC m=+105.206572323 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.422590 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dchfn\" (UniqueName: \"kubernetes.io/projected/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-kube-api-access-dchfn\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.423571 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-qp2t9" event={"ID":"482a0770-6dc5-4682-bf1e-e3c085129f33","Type":"ContainerStarted","Data":"7659d6f88727dba4939f6556562e9329b94862fcf85fcee7962b3224d40fc6f5"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.431148 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g8l7\" (UniqueName: \"kubernetes.io/projected/af2a44fd-39a9-41fc-9d1f-640e828d6a16-kube-api-access-9g8l7\") pod \"marketplace-operator-547dbd544d-ll4kg\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.437312 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6hmq\" (UniqueName: \"kubernetes.io/projected/258bf46a-4978-43d1-8579-46bfe65d8ba8-kube-api-access-h6hmq\") pod \"service-ca-74545575db-dmc74\" (UID: \"258bf46a-4978-43d1-8579-46bfe65d8ba8\") " pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.440822 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" event={"ID":"a29e146a-be7b-49f8-b97b-31f4c976afb5","Type":"ContainerStarted","Data":"d6b33be80794ace1b337dea7a890f0a2017717b2b887c92b65e9190537832921"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.444243 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" event={"ID":"beedce19-bd88-4e81-a204-99cc121cd861","Type":"ContainerStarted","Data":"1b011d84dfe35f474cea62788723e7e33d47021bf4b2bab5f89f831f0a5ceb09"} Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.444600 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-m9nhf" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.451190 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tt25\" (UniqueName: \"kubernetes.io/projected/ee00846f-ea12-4777-a333-942c9e5e7cea-kube-api-access-8tt25\") pod \"control-plane-machine-set-operator-75ffdb6fcd-qt6hv\" (UID: \"ee00846f-ea12-4777-a333-942c9e5e7cea\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.473285 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-79xmn\" (UniqueName: \"kubernetes.io/projected/384266b7-4505-4bb0-9330-ad1a2eebba3c-kube-api-access-79xmn\") pod \"service-ca-operator-5b9c976747-n7fn2\" (UID: \"384266b7-4505-4bb0-9330-ad1a2eebba3c\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.505745 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ltk9\" (UniqueName: \"kubernetes.io/projected/724bd3f7-7190-428f-8ab9-05416c2c1096-kube-api-access-7ltk9\") pod \"olm-operator-5cdf44d969-26c72\" (UID: \"724bd3f7-7190-428f-8ab9-05416c2c1096\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.515808 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.516413 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.517951 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.017911106 +0000 UTC m=+105.311198278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.518015 5117 patch_prober.go:28] interesting pod/route-controller-manager-776cdc94d6-l6ptk container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.518073 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" podUID="a29e146a-be7b-49f8-b97b-31f4c976afb5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.528862 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.529263 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.532368 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qwbc\" (UniqueName: \"kubernetes.io/projected/4db781b1-fcbc-44e9-aa0f-56184c19f397-kube-api-access-8qwbc\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.566392 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.599768 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.624857 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.627641 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.12706311 +0000 UTC m=+105.420350512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.664090 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.675506 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-74545575db-dmc74" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.687696 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.687773 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.726405 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.726641 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4db781b1-fcbc-44e9-aa0f-56184c19f397-metrics-tls\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.726676 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-node-bootstrap-token\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.726721 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.726744 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-certs\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.738524 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.238492655 +0000 UTC m=+105.531779827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.740945 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-jjwjr\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.744979 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4db781b1-fcbc-44e9-aa0f-56184c19f397-metrics-tls\") pod \"dns-default-gbnbk\" (UID: \"4db781b1-fcbc-44e9-aa0f-56184c19f397\") " pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.747464 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-certs\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.767387 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-755bb95488-d9bzl" podStartSLOduration=83.767370586 podStartE2EDuration="1m23.767370586s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:11.76584479 +0000 UTC m=+105.059131962" watchObservedRunningTime="2026-03-21 00:13:11.767370586 +0000 UTC m=+105.060657758" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.798919 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.800413 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7b571f34-0996-43b7-b450-cb7273d9ab00-node-bootstrap-token\") pod \"machine-config-server-6qjwh\" (UID: \"7b571f34-0996-43b7-b450-cb7273d9ab00\") " pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.829271 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.830225 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.330202996 +0000 UTC m=+105.623490158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.876515 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6qjwh" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.931455 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:11 crc kubenswrapper[5117]: E0321 00:13:11.931791 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.431765199 +0000 UTC m=+105.725052371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.946698 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:11 crc kubenswrapper[5117]: I0321 00:13:11.996966 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.016309 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-67c89758df-fj8qh" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.031502 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-64d44f6ddf-dg2qw" podStartSLOduration=84.031481824 podStartE2EDuration="1m24.031481824s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.030723371 +0000 UTC m=+105.324010533" watchObservedRunningTime="2026-03-21 00:13:12.031481824 +0000 UTC m=+105.324768996" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.035712 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.036634 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.536618821 +0000 UTC m=+105.829905993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.072952 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.138663 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.139160 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.639142572 +0000 UTC m=+105.932429744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.196719 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" podStartSLOduration=84.196697951 podStartE2EDuration="1m24.196697951s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.155885494 +0000 UTC m=+105.449172666" watchObservedRunningTime="2026-03-21 00:13:12.196697951 +0000 UTC m=+105.489985123" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.206568 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" podStartSLOduration=84.206547612 podStartE2EDuration="1m24.206547612s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.196790644 +0000 UTC m=+105.490077836" watchObservedRunningTime="2026-03-21 00:13:12.206547612 +0000 UTC m=+105.499834784" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.240550 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.240981 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.740963253 +0000 UTC m=+106.034250415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.268013 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-747b44746d-7zctx" podStartSLOduration=84.267998209 podStartE2EDuration="1m24.267998209s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.267236386 +0000 UTC m=+105.560523558" watchObservedRunningTime="2026-03-21 00:13:12.267998209 +0000 UTC m=+105.561285381" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.343422 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.344182 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.844164445 +0000 UTC m=+106.137451617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.344771 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-67c89758df-fj8qh" podStartSLOduration=84.344750864 podStartE2EDuration="1m24.344750864s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.343325351 +0000 UTC m=+105.636612523" watchObservedRunningTime="2026-03-21 00:13:12.344750864 +0000 UTC m=+105.638038036" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.346571 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6"] Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.362105 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29567520-5stkm" podStartSLOduration=84.362076163 podStartE2EDuration="1m24.362076163s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.361872347 +0000 UTC m=+105.655159529" watchObservedRunningTime="2026-03-21 00:13:12.362076163 +0000 UTC m=+105.655363335" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.387345 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2"] Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.401039 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn"] Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.403101 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2n4q9"] Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.453290 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.453369 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.454209 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.454684 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:12.954659311 +0000 UTC m=+106.247946483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.481301 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" event={"ID":"ee9b5fe1-7a59-49c5-b486-a6d83d56728c","Type":"ContainerStarted","Data":"e106ab1dce81c279867542bdfef7aad624495b6b44728187c9d85911e171c5e4"} Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.499674 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6qjwh" event={"ID":"7b571f34-0996-43b7-b450-cb7273d9ab00","Type":"ContainerStarted","Data":"4a72518aaf3af2b17f2ea45d416ffd6b9d48e166b25c762f0685fc00bf256eaf"} Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.515365 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" event={"ID":"1a398b4e-beca-4bbc-a41a-76f738f128d8","Type":"ContainerStarted","Data":"a48799fa95396479b3cd608f702d540410a2cce61c88d38fcad3271a3bc814b4"} Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.525378 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" event={"ID":"3a9dda43-9285-4470-8175-f919dd943c17","Type":"ContainerStarted","Data":"3117dd6ccb631129a1a020bc061c698c7e47a056c1ddf8d73a1fc28b419baa67"} Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.544099 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-qp2t9" event={"ID":"482a0770-6dc5-4682-bf1e-e3c085129f33","Type":"ContainerStarted","Data":"80ee2ac82b27aa2c5ba0e9b577596c37da6dc90899bac23c10ed1a7e0ac9671f"} Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.550629 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" podStartSLOduration=84.550602951 podStartE2EDuration="1m24.550602951s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.547396134 +0000 UTC m=+105.840683316" watchObservedRunningTime="2026-03-21 00:13:12.550602951 +0000 UTC m=+105.843890123" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.555626 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.556001 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.055963186 +0000 UTC m=+106.349250358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.625050 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-5zkpw" podStartSLOduration=84.625027875 podStartE2EDuration="1m24.625027875s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.591432899 +0000 UTC m=+105.884720071" watchObservedRunningTime="2026-03-21 00:13:12.625027875 +0000 UTC m=+105.918315187" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.625157 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-54c688565-wrfjw" podStartSLOduration=84.625153239 podStartE2EDuration="1m24.625153239s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.624644883 +0000 UTC m=+105.917932065" watchObservedRunningTime="2026-03-21 00:13:12.625153239 +0000 UTC m=+105.918440411" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.629152 5117 patch_prober.go:28] interesting pod/downloads-747b44746d-7zctx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.629260 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-747b44746d-7zctx" podUID="6c43692a-55c8-4d6a-bd36-9c3268722421" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.646017 5117 patch_prober.go:28] interesting pod/oauth-openshift-66458b6674-mtrnq container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.646075 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" podUID="7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.658195 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.659067 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.159018524 +0000 UTC m=+106.452305696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.669151 5117 patch_prober.go:28] interesting pod/route-controller-manager-776cdc94d6-l6ptk container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.669721 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" podUID="a29e146a-be7b-49f8-b97b-31f4c976afb5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.759292 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.766983 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.26694206 +0000 UTC m=+106.560229382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.806185 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-89dmq" podStartSLOduration=84.806158528 podStartE2EDuration="1m24.806158528s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:12.773964895 +0000 UTC m=+106.067252087" watchObservedRunningTime="2026-03-21 00:13:12.806158528 +0000 UTC m=+106.099445700" Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.862858 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.863225 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.363211881 +0000 UTC m=+106.656499053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.889139 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx"] Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.942763 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k"] Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.951054 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8"] Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.955090 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6"] Mar 21 00:13:12 crc kubenswrapper[5117]: I0321 00:13:12.964217 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:12 crc kubenswrapper[5117]: E0321 00:13:12.965338 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.465311949 +0000 UTC m=+106.758599121 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.069237 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:13 crc kubenswrapper[5117]: E0321 00:13:13.069674 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.569659037 +0000 UTC m=+106.862946209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.176911 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:13 crc kubenswrapper[5117]: E0321 00:13:13.177995 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.677964405 +0000 UTC m=+106.971251577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.181833 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.184762 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.188660 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.212124 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.217734 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" podStartSLOduration=85.21771454 podStartE2EDuration="1m25.21771454s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:13.198174373 +0000 UTC m=+106.491461555" watchObservedRunningTime="2026-03-21 00:13:13.21771454 +0000 UTC m=+106.511001712" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.224599 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-ll4kg"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.283704 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:13 crc kubenswrapper[5117]: E0321 00:13:13.284195 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.78418137 +0000 UTC m=+107.077468542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.297698 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.297788 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.327179 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.332673 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.356149 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-69db94689b-twxrl"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.376622 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:13 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:13 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:13 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.376697 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.384953 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:13 crc kubenswrapper[5117]: E0321 00:13:13.385342 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.885323159 +0000 UTC m=+107.178610331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.418337 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-74545575db-dmc74"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.428258 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-m9nhf"] Mar 21 00:13:13 crc kubenswrapper[5117]: W0321 00:13:13.446727 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d8ab52c_cb4f_467e_9044_1772b203fea8.slice/crio-0e6a5e22778f0d1b3050b1873d4b2258409ea9bc23e95658886012fd5a594fc8 WatchSource:0}: Error finding container 0e6a5e22778f0d1b3050b1873d4b2258409ea9bc23e95658886012fd5a594fc8: Status 404 returned error can't find the container with id 0e6a5e22778f0d1b3050b1873d4b2258409ea9bc23e95658886012fd5a594fc8 Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.491439 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:13 crc kubenswrapper[5117]: E0321 00:13:13.491837 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:13.991819283 +0000 UTC m=+107.285106455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.493851 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.498938 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-wzx89" podStartSLOduration=85.49890875 podStartE2EDuration="1m25.49890875s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:13.490892115 +0000 UTC m=+106.784179307" watchObservedRunningTime="2026-03-21 00:13:13.49890875 +0000 UTC m=+106.792195932" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.500100 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.505786 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.528161 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gbnbk"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.598517 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:13 crc kubenswrapper[5117]: E0321 00:13:13.599717 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.099671418 +0000 UTC m=+107.392958590 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.660078 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" event={"ID":"c5113992-e2f4-427c-acf3-55a4ef62376c","Type":"ContainerStarted","Data":"0da07b17cc4356a112825383fd52fa93d05eeecc61a1b1482afb896caaf5127c"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.663704 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podStartSLOduration=85.663681883 podStartE2EDuration="1m25.663681883s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:13.608318242 +0000 UTC m=+106.901605414" watchObservedRunningTime="2026-03-21 00:13:13.663681883 +0000 UTC m=+106.956969045" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.668002 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb"] Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.702763 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:13 crc kubenswrapper[5117]: E0321 00:13:13.703347 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.203327254 +0000 UTC m=+107.496614426 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.725954 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" event={"ID":"9d8ab52c-cb4f-467e-9044-1772b203fea8","Type":"ContainerStarted","Data":"0e6a5e22778f0d1b3050b1873d4b2258409ea9bc23e95658886012fd5a594fc8"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.750742 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" event={"ID":"7ed37e18-269e-4b50-a36a-0a723d5377d7","Type":"ContainerStarted","Data":"1def42ef0e1f42cf33069b423f0427457ad04adc663b3384c5e945d70ce4a69a"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.760001 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" event={"ID":"b91d27f8-a52a-4a5a-a913-782a11c8e051","Type":"ContainerStarted","Data":"df0c93e73c91032680523e05308bc8f5978952dced97cd1bb56ad26254b63bac"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.766939 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" event={"ID":"0250d937-6fd3-40f7-802a-1b7cf81f3d5f","Type":"ContainerStarted","Data":"d0c049990cc1b9b74bba2ac8db1a743a6c56aac2a8a13bf164e23ab0477ad4b3"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.777604 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx" event={"ID":"b77f02bf-fa45-4ac2-bf1c-aa01afc3f637","Type":"ContainerStarted","Data":"d65e9b808d46475f973a4fe3bfe9409e79ceaf54d6873f5d505da368d064133c"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.786485 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" event={"ID":"16a553dc-4b35-4c23-877a-4c7915ad94a4","Type":"ContainerStarted","Data":"df64f419793af81bf4ddfb459717b57cee190042119b806f3f41a726462f6472"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.786539 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" event={"ID":"16a553dc-4b35-4c23-877a-4c7915ad94a4","Type":"ContainerStarted","Data":"441984276d6a5b45f1ea686a97b87e1bffa68e25a956a2c10a88b58be5cbd285"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.805063 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.805594 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:13 crc kubenswrapper[5117]: E0321 00:13:13.805888 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.305868646 +0000 UTC m=+107.599155818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.815419 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" event={"ID":"af2a44fd-39a9-41fc-9d1f-640e828d6a16","Type":"ContainerStarted","Data":"0fedcbe439338cc00a1ddeb8853f641888f8c1bff6b0647aecc8b73e3e30dc05"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.815531 5117 patch_prober.go:28] interesting pod/packageserver-7d4fc7d867-k4rmn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" start-of-body= Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.815637 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" podUID="16a553dc-4b35-4c23-877a-4c7915ad94a4" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.822003 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" event={"ID":"d9e554a8-6f94-4a46-acd7-3446e17421a5","Type":"ContainerStarted","Data":"0ab9334abb284b54d4002b7db9baf66bd246cfe8ef01e97750b5989300b0e81c"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.836870 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" event={"ID":"35de0468-bf61-4986-8473-77b7de39194a","Type":"ContainerStarted","Data":"15ffc80a6bbaed417879cf6046add8309351e9ad23dfb34e26c8c37f9860010d"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.851461 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" event={"ID":"e9d6f20d-0da4-4ae3-b0b8-99887e1cc7e3","Type":"ContainerStarted","Data":"1df1111eb5ff60b66f85a8bff3adde03be5ba97378b2b83e52ed989beac3a2a2"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.856685 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" event={"ID":"1280d6d3-8430-41a6-9e76-5ddb21a0f44e","Type":"ContainerStarted","Data":"9cdd66e7389df042b25c03252bac0c60344c08edba04e161fe9c189aa0465224"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.875022 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" event={"ID":"9f6e0dec-9f57-403b-893c-5a30c576a799","Type":"ContainerStarted","Data":"26c177b9b17cae3c3c2eb5e119ac9fdabb537d28980e35f7baea846039f870a3"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.884005 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6qjwh" event={"ID":"7b571f34-0996-43b7-b450-cb7273d9ab00","Type":"ContainerStarted","Data":"9a92109ab4dedc8ba2acfb041af32daaf26359e1fa34bf2cf6dd32698466445a"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.889880 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.891487 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" event={"ID":"1a398b4e-beca-4bbc-a41a-76f738f128d8","Type":"ContainerStarted","Data":"77a494484a4f10d61024642a172c675b443b62d03ffaea56d13031589c47b18b"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.899080 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-69b85846b6-cvk4t" podStartSLOduration=85.899052103 podStartE2EDuration="1m25.899052103s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:13.895564216 +0000 UTC m=+107.188851408" watchObservedRunningTime="2026-03-21 00:13:13.899052103 +0000 UTC m=+107.192339285" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.899678 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" podStartSLOduration=85.899669582 podStartE2EDuration="1m25.899669582s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:13.833969134 +0000 UTC m=+107.127256306" watchObservedRunningTime="2026-03-21 00:13:13.899669582 +0000 UTC m=+107.192956764" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.906697 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:13 crc kubenswrapper[5117]: E0321 00:13:13.909003 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.408976295 +0000 UTC m=+107.702263627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.926861 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" podStartSLOduration=85.926846142 podStartE2EDuration="1m25.926846142s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:13.926564063 +0000 UTC m=+107.219851235" watchObservedRunningTime="2026-03-21 00:13:13.926846142 +0000 UTC m=+107.220133314" Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.949613 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" event={"ID":"f863fff9-286a-45fa-b8f0-8a86994b8440","Type":"ContainerStarted","Data":"2fbbf08600647a122bcabd537663227e7c1f5e86b08910e13099acf74d576e93"} Mar 21 00:13:13 crc kubenswrapper[5117]: I0321 00:13:13.974448 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" event={"ID":"9462f609-9733-4abc-8f7c-f5fa0041f39d","Type":"ContainerStarted","Data":"1b457205b4e36ebcda806cb01e2536c02d9910786184a06ea4b2598c1be2fec0"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.000932 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" event={"ID":"60516002-0c51-45a3-99a7-1ced5970206e","Type":"ContainerStarted","Data":"7c29176afa207a84967d995029db555b1a92b2a1f852011261052fdee812fb93"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.007981 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.008287 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.508241458 +0000 UTC m=+107.801528730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.008451 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.011070 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.511057114 +0000 UTC m=+107.804344286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.018276 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-qp2t9" event={"ID":"482a0770-6dc5-4682-bf1e-e3c085129f33","Type":"ContainerStarted","Data":"5bf043ec65a12433a05b8ec839fc7fed94de921c665591afef2fb5e5ade0001c"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.038325 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" podStartSLOduration=86.038304696 podStartE2EDuration="1m26.038304696s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:14.035262573 +0000 UTC m=+107.328549755" watchObservedRunningTime="2026-03-21 00:13:14.038304696 +0000 UTC m=+107.331591868" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.038500 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-6qjwh" podStartSLOduration=6.038495882 podStartE2EDuration="6.038495882s" podCreationTimestamp="2026-03-21 00:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:13.956353563 +0000 UTC m=+107.249640755" watchObservedRunningTime="2026-03-21 00:13:14.038495882 +0000 UTC m=+107.331783044" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.039178 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-74545575db-dmc74" event={"ID":"258bf46a-4978-43d1-8579-46bfe65d8ba8","Type":"ContainerStarted","Data":"45f8893f26412cb3972c05eb8d9db46abb143ed2b9540823d5227eda67f6d4f0"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.049639 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" event={"ID":"17b87002-b798-480a-8e17-83053d698239","Type":"ContainerStarted","Data":"389fe8b7b2d09b356263bed1bfb0e426c48d51eee48a07f47f6cc42286130a56"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.057649 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" event={"ID":"70789848-de09-49a1-8f00-e29b310194fa","Type":"ContainerStarted","Data":"647da37344ce1574b046a53b10022386ffe9bda752cffe180501f86b5e20244e"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.063440 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-qp2t9" podStartSLOduration=86.063403243 podStartE2EDuration="1m26.063403243s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:14.057998918 +0000 UTC m=+107.351286090" watchObservedRunningTime="2026-03-21 00:13:14.063403243 +0000 UTC m=+107.356690425" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.070779 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" event={"ID":"7c54d31e-9f4b-4071-8768-1eec361105d4","Type":"ContainerStarted","Data":"8e637ff959fac58363631a19413364916d5c8ee3fce1a509896ddc549814b0be"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.073593 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" event={"ID":"58bdc39b-9471-4649-b1c4-45206dd8855d","Type":"ContainerStarted","Data":"d85e504a84db307f18fc837dbe45e069db24c18a435e1104be46263843981ad0"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.082569 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" event={"ID":"bd48a05d-1bac-41bb-8b91-419cb4072895","Type":"ContainerStarted","Data":"04be4c9b85e8d372ad7703dd85a03bc9aa13c14f3dd72f13f00c8eb4f3ab7920"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.082625 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" event={"ID":"bd48a05d-1bac-41bb-8b91-419cb4072895","Type":"ContainerStarted","Data":"cd4ec64ebef6841879b77b552cf1261f91ad8470456d997524d92cd4edc194f4"} Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.084474 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.090633 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.091406 5117 patch_prober.go:28] interesting pod/catalog-operator-75ff9f647d-hbjn2 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.091462 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" podUID="bd48a05d-1bac-41bb-8b91-419cb4072895" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.101009 5117 patch_prober.go:28] interesting pod/downloads-747b44746d-7zctx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.101049 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-747b44746d-7zctx" podUID="6c43692a-55c8-4d6a-bd36-9c3268722421" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.107437 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" podStartSLOduration=86.107408747 podStartE2EDuration="1m26.107408747s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:14.105392406 +0000 UTC m=+107.398679578" watchObservedRunningTime="2026-03-21 00:13:14.107408747 +0000 UTC m=+107.400695919" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.109725 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.110430 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.610369047 +0000 UTC m=+107.903656249 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.110684 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.117624 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.617576238 +0000 UTC m=+107.910863410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.177295 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52236: no serving certificate available for the kubelet" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.204336 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.212604 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.213405 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.713355664 +0000 UTC m=+108.006642836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.270937 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52238: no serving certificate available for the kubelet" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.317217 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.317839 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.817819954 +0000 UTC m=+108.111107126 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.370517 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:14 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:14 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:14 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.370630 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.372699 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52244: no serving certificate available for the kubelet" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.419208 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.419438 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.919391267 +0000 UTC m=+108.212678519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.420241 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.420796 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:14.920770069 +0000 UTC m=+108.214057241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.466134 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52256: no serving certificate available for the kubelet" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.521707 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.522374 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.022335012 +0000 UTC m=+108.315622184 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.523557 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.524304 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.024286961 +0000 UTC m=+108.317574133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.539889 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52270: no serving certificate available for the kubelet" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.571906 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52274: no serving certificate available for the kubelet" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.624749 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.625053 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.125010779 +0000 UTC m=+108.418297961 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.625415 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.626077 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.126066641 +0000 UTC m=+108.419353813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.656209 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52286: no serving certificate available for the kubelet" Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.727065 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.727213 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.22718807 +0000 UTC m=+108.520475242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.727572 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.728362 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.228353616 +0000 UTC m=+108.521640788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.834775 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.835230 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.335207709 +0000 UTC m=+108.628494881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:14 crc kubenswrapper[5117]: I0321 00:13:14.938244 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:14 crc kubenswrapper[5117]: E0321 00:13:14.939891 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.439873406 +0000 UTC m=+108.733160578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.021554 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52302: no serving certificate available for the kubelet" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.039219 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.039449 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.539406876 +0000 UTC m=+108.832694048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.040120 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.040634 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.540612864 +0000 UTC m=+108.833900036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.106952 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx" event={"ID":"b77f02bf-fa45-4ac2-bf1c-aa01afc3f637","Type":"ContainerStarted","Data":"3aabbd90557998bbeeb73adfbcb8a39767432b0fc0fe56f055dd40430bde733f"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.107015 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx" event={"ID":"b77f02bf-fa45-4ac2-bf1c-aa01afc3f637","Type":"ContainerStarted","Data":"3a599f1500a36fbf8ac433d1b8677e53a72aba9342fb52112af717125faab18e"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.111328 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" event={"ID":"af2a44fd-39a9-41fc-9d1f-640e828d6a16","Type":"ContainerStarted","Data":"6b151359ff5811031ea25e18111f24eb4c41c6ba56051a143b4e52c08a57af35"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.112718 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.113800 5117 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-ll4kg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.113879 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.121726 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" event={"ID":"1280d6d3-8430-41a6-9e76-5ddb21a0f44e","Type":"ContainerStarted","Data":"cf1d01a351be1ac96f6c1d58b47d216377ad7be216278b99628aed3358222965"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.121769 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" event={"ID":"1280d6d3-8430-41a6-9e76-5ddb21a0f44e","Type":"ContainerStarted","Data":"2eaf11ca319c229d6093363925daa79787818507f11710f329d82dc531ca6844"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.123890 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" event={"ID":"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141","Type":"ContainerStarted","Data":"22ffa243641512a250330f2517ca9976b3c49f3c39053c237e045b614b84964b"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.124024 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" event={"ID":"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141","Type":"ContainerStarted","Data":"1453ded21796b192425a1831ee64b382f1c85ac24f88db72936c88afe4e01ff2"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.134751 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-7wxvx" podStartSLOduration=87.134734478 podStartE2EDuration="1m27.134734478s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.131598974 +0000 UTC m=+108.424886146" watchObservedRunningTime="2026-03-21 00:13:15.134734478 +0000 UTC m=+108.428021650" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.139377 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" event={"ID":"b91d27f8-a52a-4a5a-a913-782a11c8e051","Type":"ContainerStarted","Data":"d941518def6a0669be92abcb29d634ba563b766caf5835d9af47fedc62e0ff14"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.141674 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.141921 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.641901508 +0000 UTC m=+108.935188680 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.144344 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" event={"ID":"ee9b5fe1-7a59-49c5-b486-a6d83d56728c","Type":"ContainerStarted","Data":"1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.147396 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.159630 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" podStartSLOduration=87.159607179 podStartE2EDuration="1m27.159607179s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.158948509 +0000 UTC m=+108.452235681" watchObservedRunningTime="2026-03-21 00:13:15.159607179 +0000 UTC m=+108.452894341" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.186035 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-cpsp6" podStartSLOduration=87.186017935 podStartE2EDuration="1m27.186017935s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.184669115 +0000 UTC m=+108.477956287" watchObservedRunningTime="2026-03-21 00:13:15.186017935 +0000 UTC m=+108.479305107" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.193025 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" event={"ID":"1a398b4e-beca-4bbc-a41a-76f738f128d8","Type":"ContainerStarted","Data":"e76c183d5c1d47e65545a3cbcefbc9d2b44e02c0810bae8d16b285708b793f63"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.209109 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" event={"ID":"f863fff9-286a-45fa-b8f0-8a86994b8440","Type":"ContainerStarted","Data":"c2451221c5b887fb3f757b9c4e73cf1741ceced316a0b54b0ce615336571cd74"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.224404 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" podStartSLOduration=7.224381187 podStartE2EDuration="7.224381187s" podCreationTimestamp="2026-03-21 00:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.223223322 +0000 UTC m=+108.516510494" watchObservedRunningTime="2026-03-21 00:13:15.224381187 +0000 UTC m=+108.517668359" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.244751 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.245090 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.745068989 +0000 UTC m=+109.038356161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.260972 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-lkslg" podStartSLOduration=87.260939594 podStartE2EDuration="1m27.260939594s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.257587532 +0000 UTC m=+108.550874704" watchObservedRunningTime="2026-03-21 00:13:15.260939594 +0000 UTC m=+108.554226766" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.262695 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.283616 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" event={"ID":"58bdc39b-9471-4649-b1c4-45206dd8855d","Type":"ContainerStarted","Data":"34e6549bf8b7760d112a8f4c63c530230f334fade6370ffda53bb648d9b0477c"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.350900 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.352523 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.852502401 +0000 UTC m=+109.145789573 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.363680 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" event={"ID":"60516002-0c51-45a3-99a7-1ced5970206e","Type":"ContainerStarted","Data":"f8a0c434152a3c49f74ce12f7c60bb556edcd8f690eccab5ea29cabe3a441595"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.379221 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:15 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:15 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:15 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.379326 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.406780 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" event={"ID":"ee00846f-ea12-4777-a333-942c9e5e7cea","Type":"ContainerStarted","Data":"89bf1c57b6b0734e19f8f4e0eca68bee1053d1ef1852176e74456229cac7c08f"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.406902 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" event={"ID":"ee00846f-ea12-4777-a333-942c9e5e7cea","Type":"ContainerStarted","Data":"d4f5f1bc1c2ea260b0f450869ab47831be80dad595782b3391ec9fce013bbd9a"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.415967 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zgbp6" podStartSLOduration=87.415944119 podStartE2EDuration="1m27.415944119s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.389869622 +0000 UTC m=+108.683156794" watchObservedRunningTime="2026-03-21 00:13:15.415944119 +0000 UTC m=+108.709231291" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.418215 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-4xkzx" podStartSLOduration=87.418210338 podStartE2EDuration="1m27.418210338s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.411775691 +0000 UTC m=+108.705062873" watchObservedRunningTime="2026-03-21 00:13:15.418210338 +0000 UTC m=+108.711497510" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.429510 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" event={"ID":"17b87002-b798-480a-8e17-83053d698239","Type":"ContainerStarted","Data":"c505749b909aca89abe8b7561cb7c7bdfdb7e5f789ecca7289807c541ff0d1ca"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.430381 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.454098 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" event={"ID":"9d8ab52c-cb4f-467e-9044-1772b203fea8","Type":"ContainerStarted","Data":"e3f58be389336108438adafbfaad252dcab97072c5d183bedaa55c845f36ee60"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.455548 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.457175 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:15.957159748 +0000 UTC m=+109.250446920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.519078 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" event={"ID":"dde4d157-ff00-4654-9b1e-f11894b4c047","Type":"ContainerStarted","Data":"547792fc7072f319850f6450e95b636107d390b1a454eaa082abc4f030226b86"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.519157 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" event={"ID":"dde4d157-ff00-4654-9b1e-f11894b4c047","Type":"ContainerStarted","Data":"7d4b26720feb376c51e280f25d1d213f27ca0562cb4a16dde41ef6454b7d2a53"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.530067 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-m9nhf" event={"ID":"78edefc3-3169-4348-b76a-c255508c077f","Type":"ContainerStarted","Data":"5ce14e5250adf84adb5f99f0348da8454f516ff9f1e3da5120b7552816661f95"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.530120 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-m9nhf" event={"ID":"78edefc3-3169-4348-b76a-c255508c077f","Type":"ContainerStarted","Data":"c97d0f488360058f5ed7e9e8fb8adc225f59b950d58f6c3bac6f290a6d845075"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.538822 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" event={"ID":"384266b7-4505-4bb0-9330-ad1a2eebba3c","Type":"ContainerStarted","Data":"157d1d48d75ab088724122401f57681ebefef190515d255f0d2cb1c6946ab43b"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.539198 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" event={"ID":"384266b7-4505-4bb0-9330-ad1a2eebba3c","Type":"ContainerStarted","Data":"8eba41643ef1161bc3b2441c2942791481fdd1ed36f314a878b1a85e10e5b534"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.558124 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-qt6hv" podStartSLOduration=87.558104551 podStartE2EDuration="1m27.558104551s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.49486397 +0000 UTC m=+108.788151142" watchObservedRunningTime="2026-03-21 00:13:15.558104551 +0000 UTC m=+108.851391723" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.559226 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.560107 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.060077401 +0000 UTC m=+109.353364573 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.597591 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-m9nhf" podStartSLOduration=8.597559477 podStartE2EDuration="8.597559477s" podCreationTimestamp="2026-03-21 00:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.57866513 +0000 UTC m=+108.871952302" watchObservedRunningTime="2026-03-21 00:13:15.597559477 +0000 UTC m=+108.890846649" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.656036 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" event={"ID":"d9e554a8-6f94-4a46-acd7-3446e17421a5","Type":"ContainerStarted","Data":"c12aae06be24a7dfff8dcbf4281c1af5a2b1212cb07e9a38c0347d800a407c07"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.656392 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" event={"ID":"35de0468-bf61-4986-8473-77b7de39194a","Type":"ContainerStarted","Data":"41ecc4540eb3d977053de5024addbdd5e4ebf634df8122a0803b6e2232d47922"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.658362 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gbnbk" event={"ID":"4db781b1-fcbc-44e9-aa0f-56184c19f397","Type":"ContainerStarted","Data":"47842d516809f38caa1c7bce40a24b25d54c6200cd048848d91a9c93a7131922"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.682116 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.684486 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.184465231 +0000 UTC m=+109.477752403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.694868 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" event={"ID":"724bd3f7-7190-428f-8ab9-05416c2c1096","Type":"ContainerStarted","Data":"0c20d32ac87410e70e9fae65d8750462e4bfe46b24d415831a6d5c3afb9442b2"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.694934 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" event={"ID":"724bd3f7-7190-428f-8ab9-05416c2c1096","Type":"ContainerStarted","Data":"1f18ad5e0ed3a7412b3efa6c4cfa98cf6a6f952f3ff06da475ed0f89e4d361d8"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.697193 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.707458 5117 patch_prober.go:28] interesting pod/olm-operator-5cdf44d969-26c72 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.707870 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" podUID="724bd3f7-7190-428f-8ab9-05416c2c1096" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.717968 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-x4xb5" podStartSLOduration=87.717931814 podStartE2EDuration="1m27.717931814s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.659072806 +0000 UTC m=+108.952359978" watchObservedRunningTime="2026-03-21 00:13:15.717931814 +0000 UTC m=+109.011218976" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.734107 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-74545575db-dmc74" event={"ID":"258bf46a-4978-43d1-8579-46bfe65d8ba8","Type":"ContainerStarted","Data":"764e2355927fefd83d4ba25a6d784e99058d02f47dce6b285a6cab2883ee86d4"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.739822 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52318: no serving certificate available for the kubelet" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.753086 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" podStartSLOduration=87.753055397 podStartE2EDuration="1m27.753055397s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.708088983 +0000 UTC m=+109.001376165" watchObservedRunningTime="2026-03-21 00:13:15.753055397 +0000 UTC m=+109.046342569" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.758664 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" event={"ID":"7c54d31e-9f4b-4071-8768-1eec361105d4","Type":"ContainerStarted","Data":"18941c9184134cdc6a5620f1f335c85f4483abea64f5b8ee6ba739b23d7f8fa6"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.758740 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" event={"ID":"7c54d31e-9f4b-4071-8768-1eec361105d4","Type":"ContainerStarted","Data":"30112c1a138beb80bca8abeaf4a948ae11d6b50d56fd9f47b82a95d4cfd7991a"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.759191 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.772850 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" event={"ID":"c5113992-e2f4-427c-acf3-55a4ef62376c","Type":"ContainerStarted","Data":"227e20e8d53851b846c52e9f1b57e2942d5f16604c00e514f518b0fe9f5de5f0"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.785735 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" event={"ID":"7ed37e18-269e-4b50-a36a-0a723d5377d7","Type":"ContainerStarted","Data":"2801ef6bba7a3b679105680fb37fe946583fe0173cd4a6731846385d06837c57"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.786639 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.789128 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.289095968 +0000 UTC m=+109.582383150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.819513 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-wmwn8" podStartSLOduration=87.819492456 podStartE2EDuration="1m27.819492456s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.753800739 +0000 UTC m=+109.047087911" watchObservedRunningTime="2026-03-21 00:13:15.819492456 +0000 UTC m=+109.112779628" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.821425 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-cn4pf" podStartSLOduration=87.821419685 podStartE2EDuration="1m27.821419685s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.811913255 +0000 UTC m=+109.105200437" watchObservedRunningTime="2026-03-21 00:13:15.821419685 +0000 UTC m=+109.114706847" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.860210 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" event={"ID":"0250d937-6fd3-40f7-802a-1b7cf81f3d5f","Type":"ContainerStarted","Data":"38c11db88a453c5fb6e32093d1c56746c08b2bb666f224f933b8d5a39f02b026"} Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.878795 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-8596bd845d-nd9tz" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.887970 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.888407 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.388391661 +0000 UTC m=+109.681678833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.893668 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-hbjn2" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.907768 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-n7fn2" podStartSLOduration=87.907743282 podStartE2EDuration="1m27.907743282s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.845649415 +0000 UTC m=+109.138936587" watchObservedRunningTime="2026-03-21 00:13:15.907743282 +0000 UTC m=+109.201030454" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.909713 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-5777786469-xqfxc" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.910667 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" podStartSLOduration=87.910659001 podStartE2EDuration="1m27.910659001s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.893080154 +0000 UTC m=+109.186367326" watchObservedRunningTime="2026-03-21 00:13:15.910659001 +0000 UTC m=+109.203946173" Mar 21 00:13:15 crc kubenswrapper[5117]: I0321 00:13:15.991223 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:15 crc kubenswrapper[5117]: E0321 00:13:15.993066 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.493043288 +0000 UTC m=+109.786330460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.046539 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-nj24k" podStartSLOduration=88.046511851 podStartE2EDuration="1m28.046511851s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:15.999939778 +0000 UTC m=+109.293226950" watchObservedRunningTime="2026-03-21 00:13:16.046511851 +0000 UTC m=+109.339799023" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.078933 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" podStartSLOduration=88.07891202 podStartE2EDuration="1m28.07891202s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:16.045226271 +0000 UTC m=+109.338513453" watchObservedRunningTime="2026-03-21 00:13:16.07891202 +0000 UTC m=+109.372199192" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.094681 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.095199 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.595174127 +0000 UTC m=+109.888461299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.110685 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" podStartSLOduration=88.11066419 podStartE2EDuration="1m28.11066419s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:16.078049484 +0000 UTC m=+109.371336656" watchObservedRunningTime="2026-03-21 00:13:16.11066419 +0000 UTC m=+109.403951362" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.112192 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-74545575db-dmc74" podStartSLOduration=88.112185576 podStartE2EDuration="1m28.112185576s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:16.109752532 +0000 UTC m=+109.403039704" watchObservedRunningTime="2026-03-21 00:13:16.112185576 +0000 UTC m=+109.405472748" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.150403 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-799b87ffcd-p2jqd" podStartSLOduration=88.150378504 podStartE2EDuration="1m28.150378504s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:16.147777385 +0000 UTC m=+109.441064557" watchObservedRunningTime="2026-03-21 00:13:16.150378504 +0000 UTC m=+109.443665676" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.196241 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.196410 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.696373199 +0000 UTC m=+109.989660371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.197007 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.197441 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.697426311 +0000 UTC m=+109.990713483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.232374 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-jjwjr"] Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.298855 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.299306 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.799241691 +0000 UTC m=+110.092528863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.299791 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.300232 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.800223861 +0000 UTC m=+110.093511033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.359036 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:16 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:16 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:16 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.359133 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.401574 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.401788 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.901747603 +0000 UTC m=+110.195034775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.402206 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.402747 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:16.902724643 +0000 UTC m=+110.196011815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.407560 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-k4rmn" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.503743 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.504024 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.003993776 +0000 UTC m=+110.297280948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.605800 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.608325 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.108307212 +0000 UTC m=+110.401594384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.707627 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.707886 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.207835583 +0000 UTC m=+110.501122755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.708611 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.709067 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.20905739 +0000 UTC m=+110.502344562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.810160 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.810449 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.310400436 +0000 UTC m=+110.603687608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.810606 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.811204 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.3111714 +0000 UTC m=+110.604458612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.873029 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gbnbk" event={"ID":"4db781b1-fcbc-44e9-aa0f-56184c19f397","Type":"ContainerStarted","Data":"b90639a92bc3139467da6f30fb31690ceae3387850002b1699c9de31b565b352"} Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.873099 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gbnbk" event={"ID":"4db781b1-fcbc-44e9-aa0f-56184c19f397","Type":"ContainerStarted","Data":"f72ce35860d8f80b6ed13a926142059d460970cf9016490f603c0046919ec964"} Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.873529 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.879370 5117 generic.go:358] "Generic (PLEG): container finished" podID="0250d937-6fd3-40f7-802a-1b7cf81f3d5f" containerID="38c11db88a453c5fb6e32093d1c56746c08b2bb666f224f933b8d5a39f02b026" exitCode=0 Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.879481 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" event={"ID":"0250d937-6fd3-40f7-802a-1b7cf81f3d5f","Type":"ContainerDied","Data":"38c11db88a453c5fb6e32093d1c56746c08b2bb666f224f933b8d5a39f02b026"} Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.884764 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" event={"ID":"70789848-de09-49a1-8f00-e29b310194fa","Type":"ContainerStarted","Data":"87040cba2524227c56a4c9e3e4082a41b677e51669f26c6e2918a1662ea20024"} Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.884833 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" event={"ID":"70789848-de09-49a1-8f00-e29b310194fa","Type":"ContainerStarted","Data":"ee338975b59bc1e768a6f7573f885c61ea490e7a5c03fee7454e713fbc20aa6c"} Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.895238 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-l5tfb" event={"ID":"dde4d157-ff00-4654-9b1e-f11894b4c047","Type":"ContainerStarted","Data":"2bb1c79ef40c3a0b899b130eddd7974b56238480cfcc9f53c9e19e562a6e8eef"} Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.898628 5117 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-ll4kg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.898758 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.900559 5117 patch_prober.go:28] interesting pod/olm-operator-5cdf44d969-26c72 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.900638 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" podUID="724bd3f7-7190-428f-8ab9-05416c2c1096" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.912525 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.912830 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.412749332 +0000 UTC m=+110.706036514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.915633 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:16 crc kubenswrapper[5117]: E0321 00:13:16.941432 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.441410398 +0000 UTC m=+110.734697570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:16 crc kubenswrapper[5117]: I0321 00:13:16.994891 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-gbnbk" podStartSLOduration=8.99485769 podStartE2EDuration="8.99485769s" podCreationTimestamp="2026-03-21 00:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:16.939393366 +0000 UTC m=+110.232680558" watchObservedRunningTime="2026-03-21 00:13:16.99485769 +0000 UTC m=+110.288144862" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.020000 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.020618 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.520592267 +0000 UTC m=+110.813879439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.068460 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-69db94689b-twxrl" podStartSLOduration=89.068429018 podStartE2EDuration="1m29.068429018s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:17.061696022 +0000 UTC m=+110.354983194" watchObservedRunningTime="2026-03-21 00:13:17.068429018 +0000 UTC m=+110.361716190" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.122489 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.122984 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.622964313 +0000 UTC m=+110.916251485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.135396 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lgn8x"] Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.139695 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.142214 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-7cl8d\"" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.154226 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52330: no serving certificate available for the kubelet" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.169277 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lgn8x"] Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.224213 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.224415 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-utilities\") pod \"certified-operators-lgn8x\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.224458 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rdvl\" (UniqueName: \"kubernetes.io/projected/1eac1109-f55b-4a75-9d83-ad0aba830baa-kube-api-access-4rdvl\") pod \"certified-operators-lgn8x\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.224519 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-catalog-content\") pod \"certified-operators-lgn8x\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.224665 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.72464753 +0000 UTC m=+111.017934692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.325663 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-utilities\") pod \"certified-operators-lgn8x\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.325736 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4rdvl\" (UniqueName: \"kubernetes.io/projected/1eac1109-f55b-4a75-9d83-ad0aba830baa-kube-api-access-4rdvl\") pod \"certified-operators-lgn8x\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.325765 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-catalog-content\") pod \"certified-operators-lgn8x\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.325831 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.326204 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.826186422 +0000 UTC m=+111.119473594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.326937 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-utilities\") pod \"certified-operators-lgn8x\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.326981 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-catalog-content\") pod \"certified-operators-lgn8x\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.350009 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pqkbn"] Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.361109 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.364548 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:17 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:17 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:17 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.364633 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.364831 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"community-operators-dockercfg-vrd5f\"" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.367220 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rdvl\" (UniqueName: \"kubernetes.io/projected/1eac1109-f55b-4a75-9d83-ad0aba830baa-kube-api-access-4rdvl\") pod \"certified-operators-lgn8x\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.375223 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pqkbn"] Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.427328 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.427632 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-catalog-content\") pod \"community-operators-pqkbn\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.427683 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-utilities\") pod \"community-operators-pqkbn\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.427707 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhfmx\" (UniqueName: \"kubernetes.io/projected/a106308f-08b0-4f04-a431-8aa7042e4fbb-kube-api-access-xhfmx\") pod \"community-operators-pqkbn\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.427809 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:17.927788125 +0000 UTC m=+111.221075297 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.469212 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.528590 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-utilities\") pod \"community-operators-pqkbn\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.528645 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xhfmx\" (UniqueName: \"kubernetes.io/projected/a106308f-08b0-4f04-a431-8aa7042e4fbb-kube-api-access-xhfmx\") pod \"community-operators-pqkbn\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.528709 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-catalog-content\") pod \"community-operators-pqkbn\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.528749 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.529144 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.02912923 +0000 UTC m=+111.322416392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.529752 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-utilities\") pod \"community-operators-pqkbn\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.530002 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-catalog-content\") pod \"community-operators-pqkbn\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.539362 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h696r"] Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.559662 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.568076 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhfmx\" (UniqueName: \"kubernetes.io/projected/a106308f-08b0-4f04-a431-8aa7042e4fbb-kube-api-access-xhfmx\") pod \"community-operators-pqkbn\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.613295 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h696r"] Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.630338 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.630508 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-catalog-content\") pod \"certified-operators-h696r\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.630533 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wndv\" (UniqueName: \"kubernetes.io/projected/4b71d84d-21ab-478c-9d5e-4ca7c422b358-kube-api-access-4wndv\") pod \"certified-operators-h696r\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.630563 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-utilities\") pod \"certified-operators-h696r\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.630744 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.130715924 +0000 UTC m=+111.424003096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.696711 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.732180 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-catalog-content\") pod \"certified-operators-h696r\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.732737 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4wndv\" (UniqueName: \"kubernetes.io/projected/4b71d84d-21ab-478c-9d5e-4ca7c422b358-kube-api-access-4wndv\") pod \"certified-operators-h696r\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.732693 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-catalog-content\") pod \"certified-operators-h696r\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.732811 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-utilities\") pod \"certified-operators-h696r\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.733220 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.733484 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-utilities\") pod \"certified-operators-h696r\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.733854 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.233836684 +0000 UTC m=+111.527123856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.743417 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mz5qp"] Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.758127 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.780399 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wndv\" (UniqueName: \"kubernetes.io/projected/4b71d84d-21ab-478c-9d5e-4ca7c422b358-kube-api-access-4wndv\") pod \"certified-operators-h696r\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.783373 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mz5qp"] Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.836070 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.837030 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrlnz\" (UniqueName: \"kubernetes.io/projected/30a52c1e-3bfa-4789-a6ab-d851d6139934-kube-api-access-lrlnz\") pod \"community-operators-mz5qp\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.837136 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-catalog-content\") pod \"community-operators-mz5qp\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.837187 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-utilities\") pod \"community-operators-mz5qp\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.837342 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.337317025 +0000 UTC m=+111.630604197 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.883230 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.938724 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.938771 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-catalog-content\") pod \"community-operators-mz5qp\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.938818 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-utilities\") pod \"community-operators-mz5qp\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.938843 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lrlnz\" (UniqueName: \"kubernetes.io/projected/30a52c1e-3bfa-4789-a6ab-d851d6139934-kube-api-access-lrlnz\") pod \"community-operators-mz5qp\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:17 crc kubenswrapper[5117]: E0321 00:13:17.939463 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.439448415 +0000 UTC m=+111.732735587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.939944 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-catalog-content\") pod \"community-operators-mz5qp\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.940156 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-utilities\") pod \"community-operators-mz5qp\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.945491 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" event={"ID":"9462f609-9733-4abc-8f7c-f5fa0041f39d","Type":"ContainerStarted","Data":"0d9ccc4c7c038ce59735503b21bf8afe2a8d201d226870c234a34f461b764158"} Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.952599 5117 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-ll4kg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.952695 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.958886 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" podUID="ee9b5fe1-7a59-49c5-b486-a6d83d56728c" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" gracePeriod=30 Mar 21 00:13:17 crc kubenswrapper[5117]: I0321 00:13:17.985740 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-26c72" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:17.999981 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrlnz\" (UniqueName: \"kubernetes.io/projected/30a52c1e-3bfa-4789-a6ab-d851d6139934-kube-api-access-lrlnz\") pod \"community-operators-mz5qp\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.003563 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lgn8x"] Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.045987 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.047690 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.54765928 +0000 UTC m=+111.840946452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.124041 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.151001 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.151646 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.651631026 +0000 UTC m=+111.944918198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.254120 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.254556 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.75453493 +0000 UTC m=+112.047822102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.356508 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.357441 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.857423012 +0000 UTC m=+112.150710184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.359568 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:18 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:18 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:18 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.359652 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.457905 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.458279 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:18.958244282 +0000 UTC m=+112.251531454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.465580 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.473334 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h696r"] Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.539894 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pqkbn"] Mar 21 00:13:18 crc kubenswrapper[5117]: W0321 00:13:18.552416 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda106308f_08b0_4f04_a431_8aa7042e4fbb.slice/crio-4a9cf04658d3d2f0b8eef892a4b5461f0d31740e15bc853103f6f9bf17def73b WatchSource:0}: Error finding container 4a9cf04658d3d2f0b8eef892a4b5461f0d31740e15bc853103f6f9bf17def73b: Status 404 returned error can't find the container with id 4a9cf04658d3d2f0b8eef892a4b5461f0d31740e15bc853103f6f9bf17def73b Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.576518 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-secret-volume\") pod \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.576577 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-config-volume\") pod \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.576809 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvhtm\" (UniqueName: \"kubernetes.io/projected/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-kube-api-access-pvhtm\") pod \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\" (UID: \"0250d937-6fd3-40f7-802a-1b7cf81f3d5f\") " Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.576965 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.577402 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.077388032 +0000 UTC m=+112.370675204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.579340 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-config-volume" (OuterVolumeSpecName: "config-volume") pod "0250d937-6fd3-40f7-802a-1b7cf81f3d5f" (UID: "0250d937-6fd3-40f7-802a-1b7cf81f3d5f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.579437 5117 scope.go:117] "RemoveContainer" containerID="79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17" Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.579626 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.597984 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0250d937-6fd3-40f7-802a-1b7cf81f3d5f" (UID: "0250d937-6fd3-40f7-802a-1b7cf81f3d5f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.619607 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mz5qp"] Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.632705 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-kube-api-access-pvhtm" (OuterVolumeSpecName: "kube-api-access-pvhtm") pod "0250d937-6fd3-40f7-802a-1b7cf81f3d5f" (UID: "0250d937-6fd3-40f7-802a-1b7cf81f3d5f"). InnerVolumeSpecName "kube-api-access-pvhtm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:18 crc kubenswrapper[5117]: W0321 00:13:18.657132 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30a52c1e_3bfa_4789_a6ab_d851d6139934.slice/crio-bda23d6d7e0f51eba0cc6fc3de72270405282e5fae544c3a9ad2d92125979633 WatchSource:0}: Error finding container bda23d6d7e0f51eba0cc6fc3de72270405282e5fae544c3a9ad2d92125979633: Status 404 returned error can't find the container with id bda23d6d7e0f51eba0cc6fc3de72270405282e5fae544c3a9ad2d92125979633 Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.677602 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.677860 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.17781724 +0000 UTC m=+112.471104412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.678224 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.678379 5117 reconciler_common.go:299] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.678398 5117 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.678424 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pvhtm\" (UniqueName: \"kubernetes.io/projected/0250d937-6fd3-40f7-802a-1b7cf81f3d5f-kube-api-access-pvhtm\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.678758 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.178750909 +0000 UTC m=+112.472038081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.681361 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.681436 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.695442 5117 patch_prober.go:28] interesting pod/apiserver-9ddfb9f55-nsfg8 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]log ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]etcd ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/generic-apiserver-start-informers ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/max-in-flight-filter ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 21 00:13:18 crc kubenswrapper[5117]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 21 00:13:18 crc kubenswrapper[5117]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/project.openshift.io-projectcache ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/openshift.io-startinformers ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 21 00:13:18 crc kubenswrapper[5117]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 21 00:13:18 crc kubenswrapper[5117]: livez check failed Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.696006 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" podUID="60516002-0c51-45a3-99a7-1ced5970206e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.757485 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.757573 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.759060 5117 patch_prober.go:28] interesting pod/console-64d44f6ddf-dg2qw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.759140 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-console/console-64d44f6ddf-dg2qw" podUID="fda377c1-1e51-4241-a939-a44cc536104d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.774423 5117 patch_prober.go:28] interesting pod/downloads-747b44746d-7zctx container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.774498 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-747b44746d-7zctx" podUID="6c43692a-55c8-4d6a-bd36-9c3268722421" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.779349 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.780887 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.280867108 +0000 UTC m=+112.574154270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.882438 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.883036 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.383015019 +0000 UTC m=+112.676302351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.967737 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/revision-pruner-6-crc"] Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.969981 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0250d937-6fd3-40f7-802a-1b7cf81f3d5f" containerName="collect-profiles" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.970009 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="0250d937-6fd3-40f7-802a-1b7cf81f3d5f" containerName="collect-profiles" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.970608 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="0250d937-6fd3-40f7-802a-1b7cf81f3d5f" containerName="collect-profiles" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.977093 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.979622 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler\"/\"installer-sa-dockercfg-qpkss\"" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.980723 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler\"/\"kube-root-ca.crt\"" Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.984087 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:18 crc kubenswrapper[5117]: E0321 00:13:18.984706 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.484679914 +0000 UTC m=+112.777967086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.997270 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h696r" event={"ID":"4b71d84d-21ab-478c-9d5e-4ca7c422b358","Type":"ContainerStarted","Data":"feaab3061bfc5e8f9424e93af4c9a1af21b957460251a8a8d64aba5c6376d387"} Mar 21 00:13:18 crc kubenswrapper[5117]: I0321 00:13:18.997342 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h696r" event={"ID":"4b71d84d-21ab-478c-9d5e-4ca7c422b358","Type":"ContainerStarted","Data":"07062a0dfcfa62fc9b1b51cc65647063fd403489ca12603b798dbdc17e77cc28"} Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.004611 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/revision-pruner-6-crc"] Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.020676 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.022768 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567520-pnrmd" event={"ID":"0250d937-6fd3-40f7-802a-1b7cf81f3d5f","Type":"ContainerDied","Data":"d0c049990cc1b9b74bba2ac8db1a743a6c56aac2a8a13bf164e23ab0477ad4b3"} Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.022811 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0c049990cc1b9b74bba2ac8db1a743a6c56aac2a8a13bf164e23ab0477ad4b3" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.067807 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mz5qp" event={"ID":"30a52c1e-3bfa-4789-a6ab-d851d6139934","Type":"ContainerStarted","Data":"bda23d6d7e0f51eba0cc6fc3de72270405282e5fae544c3a9ad2d92125979633"} Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.089403 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.089459 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kube-api-access\") pod \"revision-pruner-6-crc\" (UID: \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.089492 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kubelet-dir\") pod \"revision-pruner-6-crc\" (UID: \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.090924 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.590907218 +0000 UTC m=+112.884194380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.114550 5117 generic.go:358] "Generic (PLEG): container finished" podID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerID="5b22728c35965a85542c89a237d68e3ff3492ac6f1a71e55c83a7d7eca7a408b" exitCode=0 Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.114987 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lgn8x" event={"ID":"1eac1109-f55b-4a75-9d83-ad0aba830baa","Type":"ContainerDied","Data":"5b22728c35965a85542c89a237d68e3ff3492ac6f1a71e55c83a7d7eca7a408b"} Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.115020 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lgn8x" event={"ID":"1eac1109-f55b-4a75-9d83-ad0aba830baa","Type":"ContainerStarted","Data":"39b8f7e2f12cba8583afe76ba7a3d20856e9ebebeac3a689cb926030853aedf7"} Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.143323 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqkbn" event={"ID":"a106308f-08b0-4f04-a431-8aa7042e4fbb","Type":"ContainerStarted","Data":"20a499d64717d049e9c7ecb04ace9ed533bc0605d630efbd78ca9661844d2ed9"} Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.143426 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqkbn" event={"ID":"a106308f-08b0-4f04-a431-8aa7042e4fbb","Type":"ContainerStarted","Data":"4a9cf04658d3d2f0b8eef892a4b5461f0d31740e15bc853103f6f9bf17def73b"} Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.193037 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.193199 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kube-api-access\") pod \"revision-pruner-6-crc\" (UID: \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.193237 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kubelet-dir\") pod \"revision-pruner-6-crc\" (UID: \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.193810 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.693789502 +0000 UTC m=+112.987076674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.194197 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kubelet-dir\") pod \"revision-pruner-6-crc\" (UID: \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.234706 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kube-api-access\") pod \"revision-pruner-6-crc\" (UID: \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.321529 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.321949 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.821934346 +0000 UTC m=+113.115221518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.322153 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.357647 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:19 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:19 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:19 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.357731 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.371805 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bn6n6"] Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.391604 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.398022 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-gg4w7\"" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.403774 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn6n6"] Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.423429 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.423841 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:19.923819008 +0000 UTC m=+113.217106180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.525523 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-catalog-content\") pod \"redhat-marketplace-bn6n6\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.525584 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-utilities\") pod \"redhat-marketplace-bn6n6\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.525624 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.525700 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfq24\" (UniqueName: \"kubernetes.io/projected/0c19bd60-8272-4e5e-81d3-10660df3c5ed-kube-api-access-dfq24\") pod \"redhat-marketplace-bn6n6\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.526135 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.026117174 +0000 UTC m=+113.319404346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.627448 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.627660 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.127622375 +0000 UTC m=+113.420909547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.628120 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-catalog-content\") pod \"redhat-marketplace-bn6n6\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.628296 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-utilities\") pod \"redhat-marketplace-bn6n6\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.628410 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.628579 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dfq24\" (UniqueName: \"kubernetes.io/projected/0c19bd60-8272-4e5e-81d3-10660df3c5ed-kube-api-access-dfq24\") pod \"redhat-marketplace-bn6n6\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.628867 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-catalog-content\") pod \"redhat-marketplace-bn6n6\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.629127 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-utilities\") pod \"redhat-marketplace-bn6n6\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.629326 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.129307026 +0000 UTC m=+113.422594198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.663546 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfq24\" (UniqueName: \"kubernetes.io/projected/0c19bd60-8272-4e5e-81d3-10660df3c5ed-kube-api-access-dfq24\") pod \"redhat-marketplace-bn6n6\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.710469 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.728675 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rz8n5"] Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.731225 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.731387 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.231350773 +0000 UTC m=+113.524637945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.731796 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.732276 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.232268041 +0000 UTC m=+113.525555213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.744766 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.749246 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz8n5"] Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.793446 5117 ???:1] "http: TLS handshake error from 192.168.126.11:52346: no serving certificate available for the kubelet" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.803912 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/revision-pruner-6-crc"] Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.834821 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.835043 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-utilities\") pod \"redhat-marketplace-rz8n5\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.835081 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjmm9\" (UniqueName: \"kubernetes.io/projected/55a3135d-1190-4674-9ceb-a8d36fc19e0b-kube-api-access-cjmm9\") pod \"redhat-marketplace-rz8n5\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.835102 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-catalog-content\") pod \"redhat-marketplace-rz8n5\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.835324 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.335300558 +0000 UTC m=+113.628587730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.940521 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-utilities\") pod \"redhat-marketplace-rz8n5\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.940601 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cjmm9\" (UniqueName: \"kubernetes.io/projected/55a3135d-1190-4674-9ceb-a8d36fc19e0b-kube-api-access-cjmm9\") pod \"redhat-marketplace-rz8n5\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.940649 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-catalog-content\") pod \"redhat-marketplace-rz8n5\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.940732 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.941153 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-utilities\") pod \"redhat-marketplace-rz8n5\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:19 crc kubenswrapper[5117]: E0321 00:13:19.941231 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.441201064 +0000 UTC m=+113.734488236 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.941681 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-catalog-content\") pod \"redhat-marketplace-rz8n5\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:19 crc kubenswrapper[5117]: I0321 00:13:19.978437 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjmm9\" (UniqueName: \"kubernetes.io/projected/55a3135d-1190-4674-9ceb-a8d36fc19e0b-kube-api-access-cjmm9\") pod \"redhat-marketplace-rz8n5\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.042893 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.043140 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.543090485 +0000 UTC m=+113.836377657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.043637 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.044131 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.544105737 +0000 UTC m=+113.837392909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.081152 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.182972 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.183738 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.683709421 +0000 UTC m=+113.976996593 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.217631 5117 generic.go:358] "Generic (PLEG): container finished" podID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerID="feaab3061bfc5e8f9424e93af4c9a1af21b957460251a8a8d64aba5c6376d387" exitCode=0 Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.217869 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h696r" event={"ID":"4b71d84d-21ab-478c-9d5e-4ca7c422b358","Type":"ContainerDied","Data":"feaab3061bfc5e8f9424e93af4c9a1af21b957460251a8a8d64aba5c6376d387"} Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.219149 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn6n6"] Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.226531 5117 generic.go:358] "Generic (PLEG): container finished" podID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerID="ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae" exitCode=0 Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.226813 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mz5qp" event={"ID":"30a52c1e-3bfa-4789-a6ab-d851d6139934","Type":"ContainerDied","Data":"ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae"} Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.235077 5117 generic.go:358] "Generic (PLEG): container finished" podID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerID="20a499d64717d049e9c7ecb04ace9ed533bc0605d630efbd78ca9661844d2ed9" exitCode=0 Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.235163 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqkbn" event={"ID":"a106308f-08b0-4f04-a431-8aa7042e4fbb","Type":"ContainerDied","Data":"20a499d64717d049e9c7ecb04ace9ed533bc0605d630efbd78ca9661844d2ed9"} Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.240606 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/revision-pruner-6-crc" event={"ID":"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece","Type":"ContainerStarted","Data":"7d63585c1a5a9d133060f1ab58278dd8bcd7adcb77b32a6b2fc279d7200e47cd"} Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.291850 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.292170 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.792158024 +0000 UTC m=+114.085445196 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.359748 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d2v8t"] Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.364091 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:20 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:20 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:20 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.364200 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.365128 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.366049 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.370694 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-9gxlh\"" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.381351 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d2v8t"] Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.393662 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.394173 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-catalog-content\") pod \"redhat-operators-d2v8t\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.394290 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-utilities\") pod \"redhat-operators-d2v8t\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.394425 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.894386617 +0000 UTC m=+114.187673789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.395824 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgcmd\" (UniqueName: \"kubernetes.io/projected/0c7aa7ad-faf4-41a3-82a1-c064c8109629-kube-api-access-fgcmd\") pod \"redhat-operators-d2v8t\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.496978 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.497337 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-utilities\") pod \"redhat-operators-d2v8t\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.497380 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fgcmd\" (UniqueName: \"kubernetes.io/projected/0c7aa7ad-faf4-41a3-82a1-c064c8109629-kube-api-access-fgcmd\") pod \"redhat-operators-d2v8t\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.497424 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-catalog-content\") pod \"redhat-operators-d2v8t\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.497793 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-catalog-content\") pod \"redhat-operators-d2v8t\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.497928 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:20.997896669 +0000 UTC m=+114.291184021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.498773 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-utilities\") pod \"redhat-operators-d2v8t\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.538956 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgcmd\" (UniqueName: \"kubernetes.io/projected/0c7aa7ad-faf4-41a3-82a1-c064c8109629-kube-api-access-fgcmd\") pod \"redhat-operators-d2v8t\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.602964 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.603289 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.103268067 +0000 UTC m=+114.396555239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.694012 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.704226 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.704732 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.204711146 +0000 UTC m=+114.497998318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.727943 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qz57f"] Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.745951 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.762114 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qz57f"] Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.806967 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.807870 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvw2v\" (UniqueName: \"kubernetes.io/projected/c15a6f5e-ffea-444b-98d1-3922aeb22128-kube-api-access-kvw2v\") pod \"redhat-operators-qz57f\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.807897 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-utilities\") pod \"redhat-operators-qz57f\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.807957 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-catalog-content\") pod \"redhat-operators-qz57f\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.808067 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.308047643 +0000 UTC m=+114.601334805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.821360 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz8n5"] Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.871803 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-11-crc"] Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.917562 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-catalog-content\") pod \"redhat-operators-qz57f\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.918066 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-catalog-content\") pod \"redhat-operators-qz57f\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.918108 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.918131 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kvw2v\" (UniqueName: \"kubernetes.io/projected/c15a6f5e-ffea-444b-98d1-3922aeb22128-kube-api-access-kvw2v\") pod \"redhat-operators-qz57f\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.918148 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-utilities\") pod \"redhat-operators-qz57f\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.918497 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-utilities\") pod \"redhat-operators-qz57f\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:20 crc kubenswrapper[5117]: E0321 00:13:20.918815 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.418800116 +0000 UTC m=+114.712087288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.926071 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-11-crc"] Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.926319 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.928548 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver\"/\"kube-root-ca.crt\"" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.928945 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver\"/\"installer-sa-dockercfg-bqqnb\"" Mar 21 00:13:20 crc kubenswrapper[5117]: I0321 00:13:20.948711 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvw2v\" (UniqueName: \"kubernetes.io/projected/c15a6f5e-ffea-444b-98d1-3922aeb22128-kube-api-access-kvw2v\") pod \"redhat-operators-qz57f\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.021765 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d2v8t"] Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.022218 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.023131 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.523093621 +0000 UTC m=+114.816380793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.025125 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.025300 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kubelet-dir\") pod \"revision-pruner-11-crc\" (UID: \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.025572 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kube-api-access\") pod \"revision-pruner-11-crc\" (UID: \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.026379 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.526347381 +0000 UTC m=+114.819634553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.078725 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.126540 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.126776 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.626725697 +0000 UTC m=+114.920012869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.126941 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kube-api-access\") pod \"revision-pruner-11-crc\" (UID: \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.127123 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.127195 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kubelet-dir\") pod \"revision-pruner-11-crc\" (UID: \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.127445 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kubelet-dir\") pod \"revision-pruner-11-crc\" (UID: \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.127829 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.62779815 +0000 UTC m=+114.921085492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.161581 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kube-api-access\") pod \"revision-pruner-11-crc\" (UID: \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.230291 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.230853 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.730825937 +0000 UTC m=+115.024113109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.268650 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.281443 5117 generic.go:358] "Generic (PLEG): container finished" podID="4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece" containerID="1058951f17a5fc8debe2ddc90e2fb7cfb061ead82aa532362ad96da6dc19d1ea" exitCode=0 Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.281557 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/revision-pruner-6-crc" event={"ID":"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece","Type":"ContainerDied","Data":"1058951f17a5fc8debe2ddc90e2fb7cfb061ead82aa532362ad96da6dc19d1ea"} Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.286420 5117 generic.go:358] "Generic (PLEG): container finished" podID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerID="4f1d6da7c02f7a39bc544819646368d43d4cfa5dd5aa2ccfe427a4c19d74a6aa" exitCode=0 Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.286506 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn6n6" event={"ID":"0c19bd60-8272-4e5e-81d3-10660df3c5ed","Type":"ContainerDied","Data":"4f1d6da7c02f7a39bc544819646368d43d4cfa5dd5aa2ccfe427a4c19d74a6aa"} Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.286529 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn6n6" event={"ID":"0c19bd60-8272-4e5e-81d3-10660df3c5ed","Type":"ContainerStarted","Data":"8bb2135785bcf446f3d83ca4d200e1483d2eae7d5e09afbb41c81d57f23df164"} Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.301231 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2v8t" event={"ID":"0c7aa7ad-faf4-41a3-82a1-c064c8109629","Type":"ContainerStarted","Data":"8fb837ac163667d358eb90e07958df816b88628d84ad80f6ac0aa2204898fdeb"} Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.301297 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2v8t" event={"ID":"0c7aa7ad-faf4-41a3-82a1-c064c8109629","Type":"ContainerStarted","Data":"48bc262aea466c0bd201594f8a2d9794775a3cac5a695d0cba6dd51a62b10ec7"} Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.306181 5117 generic.go:358] "Generic (PLEG): container finished" podID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerID="14d9cae71ca5afd9df48f906ff026fd51a53648a761b927c5800ea835e85ba36" exitCode=0 Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.306303 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz8n5" event={"ID":"55a3135d-1190-4674-9ceb-a8d36fc19e0b","Type":"ContainerDied","Data":"14d9cae71ca5afd9df48f906ff026fd51a53648a761b927c5800ea835e85ba36"} Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.306331 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz8n5" event={"ID":"55a3135d-1190-4674-9ceb-a8d36fc19e0b","Type":"ContainerStarted","Data":"d8f39500c5cda00616dd99e61e41be69293decf722d3e222fb4afbc7e8b30688"} Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.341819 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.342887 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.842873961 +0000 UTC m=+115.136161133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.363717 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:21 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:21 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:21 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.363807 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.416386 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qz57f"] Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.450051 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.450361 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.950308643 +0000 UTC m=+115.243595815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.450568 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.451159 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:21.951143087 +0000 UTC m=+115.244430259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.560162 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.560401 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.060352303 +0000 UTC m=+115.353639475 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.561456 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.562119 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.062106688 +0000 UTC m=+115.355393860 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.632827 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-11-crc"] Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.662983 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.663599 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.163574356 +0000 UTC m=+115.456861528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.765585 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.766150 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.26613115 +0000 UTC m=+115.559418332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.867135 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.867382 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.367342572 +0000 UTC m=+115.660629744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.867786 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.868384 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.368371512 +0000 UTC m=+115.661658684 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.972226 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.972525 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.472494943 +0000 UTC m=+115.765782125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:21 crc kubenswrapper[5117]: I0321 00:13:21.972904 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:21 crc kubenswrapper[5117]: E0321 00:13:21.973275 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.473265077 +0000 UTC m=+115.766552249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.075415 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.075600 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.575563582 +0000 UTC m=+115.868850754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.076378 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.076846 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.576825201 +0000 UTC m=+115.870112373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.178312 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.178704 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.678685221 +0000 UTC m=+115.971972393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.279950 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.280321 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.780306136 +0000 UTC m=+116.073593308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.316629 5117 generic.go:358] "Generic (PLEG): container finished" podID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerID="d1ec1e95e3c26d54fc4427b668575182a6839560c78afd57090e9f962248df2f" exitCode=0 Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.316729 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz57f" event={"ID":"c15a6f5e-ffea-444b-98d1-3922aeb22128","Type":"ContainerDied","Data":"d1ec1e95e3c26d54fc4427b668575182a6839560c78afd57090e9f962248df2f"} Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.318157 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz57f" event={"ID":"c15a6f5e-ffea-444b-98d1-3922aeb22128","Type":"ContainerStarted","Data":"b4de5e9e85dbe09789e7ae5965256793925d7d960eb4993bc87034c56df7273a"} Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.321742 5117 generic.go:358] "Generic (PLEG): container finished" podID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerID="8fb837ac163667d358eb90e07958df816b88628d84ad80f6ac0aa2204898fdeb" exitCode=0 Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.321927 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2v8t" event={"ID":"0c7aa7ad-faf4-41a3-82a1-c064c8109629","Type":"ContainerDied","Data":"8fb837ac163667d358eb90e07958df816b88628d84ad80f6ac0aa2204898fdeb"} Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.323953 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"b59d6290-e0a2-4a09-85c7-adafaa952dc9","Type":"ContainerStarted","Data":"c8f7c97c82bd89d7641b19a2f6e3b59f18c883ce6b859ba55ad47b0d26355706"} Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.359515 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:22 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:22 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:22 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.359594 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.382907 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.383778 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.883739916 +0000 UTC m=+116.177027088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.485285 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.485726 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:22.985703131 +0000 UTC m=+116.278990303 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.591949 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.592297 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.092278096 +0000 UTC m=+116.385565268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.621427 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.675046 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.694318 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kube-api-access\") pod \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\" (UID: \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\") " Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.694392 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kubelet-dir\") pod \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\" (UID: \"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece\") " Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.694714 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.695065 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece" (UID: "4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.695304 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.195225851 +0000 UTC m=+116.488513023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.710579 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece" (UID: "4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.796370 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.796589 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.296553226 +0000 UTC m=+116.589840388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.797526 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.797724 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.797738 5117 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.798213 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.298196226 +0000 UTC m=+116.591483458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.899360 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.899621 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.399585593 +0000 UTC m=+116.692872765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:22 crc kubenswrapper[5117]: I0321 00:13:22.900087 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:22 crc kubenswrapper[5117]: E0321 00:13:22.900462 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.4004476 +0000 UTC m=+116.693734782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.001825 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.002407 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.502380253 +0000 UTC m=+116.795667425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.103345 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.103801 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.603782971 +0000 UTC m=+116.897070143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.204861 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.205237 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.70521979 +0000 UTC m=+116.998506962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.310699 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.311790 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.811770084 +0000 UTC m=+117.105057256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.337211 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" event={"ID":"9462f609-9733-4abc-8f7c-f5fa0041f39d","Type":"ContainerStarted","Data":"d7a28a0a4a2d71e8dd65b75e286609b018ac925a2629ab2c02778dc2b38d140c"} Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.339661 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/revision-pruner-6-crc" event={"ID":"4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece","Type":"ContainerDied","Data":"7d63585c1a5a9d133060f1ab58278dd8bcd7adcb77b32a6b2fc279d7200e47cd"} Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.339693 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d63585c1a5a9d133060f1ab58278dd8bcd7adcb77b32a6b2fc279d7200e47cd" Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.339837 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.345876 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"b59d6290-e0a2-4a09-85c7-adafaa952dc9","Type":"ContainerStarted","Data":"9673ea37a6af2b98b2b5508486f09f4ac4e53db19cfa890038bcba80c9f95d16"} Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.356842 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:23 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:23 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:23 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.356915 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.411765 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.411985 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.911954025 +0000 UTC m=+117.205241197 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.412343 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.412740 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:23.912729018 +0000 UTC m=+117.206016200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.514129 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.514434 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:24.014414984 +0000 UTC m=+117.307702156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.615202 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.615594 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:24.115580364 +0000 UTC m=+117.408867536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.688666 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.716409 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.716659 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:24.216640372 +0000 UTC m=+117.509927544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.817513 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.818165 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:24.318095671 +0000 UTC m=+117.611382843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.847687 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-9ddfb9f55-nsfg8" Mar 21 00:13:23 crc kubenswrapper[5117]: I0321 00:13:23.926016 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:23 crc kubenswrapper[5117]: E0321 00:13:23.927306 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:24.427286746 +0000 UTC m=+117.720573918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.027996 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:24 crc kubenswrapper[5117]: E0321 00:13:24.028522 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:24.528505919 +0000 UTC m=+117.821793091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.038991 5117 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.085641 5117 patch_prober.go:28] interesting pod/downloads-747b44746d-7zctx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.086053 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-747b44746d-7zctx" podUID="6c43692a-55c8-4d6a-bd36-9c3268722421" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.129236 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:24 crc kubenswrapper[5117]: E0321 00:13:24.129433 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2026-03-21 00:13:24.6293793 +0000 UTC m=+117.922666472 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.129831 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:24 crc kubenswrapper[5117]: E0321 00:13:24.131354 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2026-03-21 00:13:24.630800373 +0000 UTC m=+117.924087565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-gjth4" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.205214 5117 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-21T00:13:24.03903407Z","UUID":"5d3341b2-1374-4f29-bd26-589360af3e6f","Handler":null,"Name":"","Endpoint":""} Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.213085 5117 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.213118 5117 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.231834 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.237635 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (OuterVolumeSpecName: "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2". PluginName "kubernetes.io/csi", VolumeGIDValue "" Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.333438 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.341985 5117 csi_attacher.go:373] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.342048 5117 operation_generator.go:557] "MountVolume.MountDevice succeeded for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b1264ac67579ad07e7e9003054d44fe40dd55285a4b2f7dc74e48be1aee0868a/globalmount\"" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.355802 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:24 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:24 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:24 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.355915 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.370216 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" event={"ID":"9462f609-9733-4abc-8f7c-f5fa0041f39d","Type":"ContainerStarted","Data":"1ee1c62ed515387f5e71c5c044ece0b1bb5c3632576a19d6e09b5e5e4cdf092e"} Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.373988 5117 generic.go:358] "Generic (PLEG): container finished" podID="b59d6290-e0a2-4a09-85c7-adafaa952dc9" containerID="9673ea37a6af2b98b2b5508486f09f4ac4e53db19cfa890038bcba80c9f95d16" exitCode=0 Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.374104 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"b59d6290-e0a2-4a09-85c7-adafaa952dc9","Type":"ContainerDied","Data":"9673ea37a6af2b98b2b5508486f09f4ac4e53db19cfa890038bcba80c9f95d16"} Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.400140 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-gjth4\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.471851 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.933752 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-gjth4"] Mar 21 00:13:24 crc kubenswrapper[5117]: I0321 00:13:24.946111 5117 ???:1] "http: TLS handshake error from 192.168.126.11:46412: no serving certificate available for the kubelet" Mar 21 00:13:25 crc kubenswrapper[5117]: E0321 00:13:25.153772 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 00:13:25 crc kubenswrapper[5117]: E0321 00:13:25.165158 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 00:13:25 crc kubenswrapper[5117]: E0321 00:13:25.167944 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 00:13:25 crc kubenswrapper[5117]: E0321 00:13:25.168047 5117 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" podUID="ee9b5fe1-7a59-49c5-b486-a6d83d56728c" containerName="kube-multus-additional-cni-plugins" probeResult="unknown" Mar 21 00:13:25 crc kubenswrapper[5117]: I0321 00:13:25.356608 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:25 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:25 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:25 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:25 crc kubenswrapper[5117]: I0321 00:13:25.356680 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:25 crc kubenswrapper[5117]: I0321 00:13:25.395453 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" event={"ID":"9462f609-9733-4abc-8f7c-f5fa0041f39d","Type":"ContainerStarted","Data":"a552303109c870e821369e5850fd25eb1ad501839575d4c95df1d6c4835e2e64"} Mar 21 00:13:25 crc kubenswrapper[5117]: I0321 00:13:25.417087 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2n4q9" podStartSLOduration=18.417067145 podStartE2EDuration="18.417067145s" podCreationTimestamp="2026-03-21 00:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:25.413051151 +0000 UTC m=+118.706338313" watchObservedRunningTime="2026-03-21 00:13:25.417067145 +0000 UTC m=+118.710354627" Mar 21 00:13:25 crc kubenswrapper[5117]: I0321 00:13:25.594919 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e9b5059-1b3e-4067-a63d-2952cbe863af" path="/var/lib/kubelet/pods/9e9b5059-1b3e-4067-a63d-2952cbe863af/volumes" Mar 21 00:13:26 crc kubenswrapper[5117]: I0321 00:13:26.354576 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:26 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:26 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:26 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:26 crc kubenswrapper[5117]: I0321 00:13:26.354653 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:26 crc kubenswrapper[5117]: I0321 00:13:26.967890 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-gbnbk" Mar 21 00:13:27 crc kubenswrapper[5117]: I0321 00:13:27.354752 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:27 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:27 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:27 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:27 crc kubenswrapper[5117]: I0321 00:13:27.354825 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:27 crc kubenswrapper[5117]: I0321 00:13:27.950776 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:13:28 crc kubenswrapper[5117]: I0321 00:13:28.355222 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:28 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:28 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:28 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:28 crc kubenswrapper[5117]: I0321 00:13:28.355539 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:28 crc kubenswrapper[5117]: I0321 00:13:28.757972 5117 patch_prober.go:28] interesting pod/console-64d44f6ddf-dg2qw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 21 00:13:28 crc kubenswrapper[5117]: I0321 00:13:28.758035 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-console/console-64d44f6ddf-dg2qw" podUID="fda377c1-1e51-4241-a939-a44cc536104d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 21 00:13:28 crc kubenswrapper[5117]: I0321 00:13:28.774374 5117 patch_prober.go:28] interesting pod/downloads-747b44746d-7zctx container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Mar 21 00:13:28 crc kubenswrapper[5117]: I0321 00:13:28.774441 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-747b44746d-7zctx" podUID="6c43692a-55c8-4d6a-bd36-9c3268722421" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Mar 21 00:13:29 crc kubenswrapper[5117]: I0321 00:13:29.357240 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:29 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:29 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:29 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:29 crc kubenswrapper[5117]: I0321 00:13:29.357359 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:30 crc kubenswrapper[5117]: I0321 00:13:30.355429 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:30 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:30 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:30 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:30 crc kubenswrapper[5117]: I0321 00:13:30.356024 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:30 crc kubenswrapper[5117]: I0321 00:13:30.423328 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" event={"ID":"f85237c3-327f-4a9e-8496-00c9087e5aa7","Type":"ContainerStarted","Data":"b9b68b63dfd72250771d9143f4fd8af7f3e22f042dbdfa6f929caeeaf4a9f3ad"} Mar 21 00:13:30 crc kubenswrapper[5117]: I0321 00:13:30.578805 5117 scope.go:117] "RemoveContainer" containerID="79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17" Mar 21 00:13:31 crc kubenswrapper[5117]: I0321 00:13:31.355232 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:31 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:31 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:31 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:31 crc kubenswrapper[5117]: I0321 00:13:31.355329 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:32 crc kubenswrapper[5117]: I0321 00:13:32.355039 5117 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-zkpbt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 21 00:13:32 crc kubenswrapper[5117]: [-]has-synced failed: reason withheld Mar 21 00:13:32 crc kubenswrapper[5117]: [+]process-running ok Mar 21 00:13:32 crc kubenswrapper[5117]: healthz check failed Mar 21 00:13:32 crc kubenswrapper[5117]: I0321 00:13:32.355186 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" podUID="55cf9bde-83fe-4510-a1ae-acaf82c35b62" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 21 00:13:33 crc kubenswrapper[5117]: I0321 00:13:33.356314 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:33 crc kubenswrapper[5117]: I0321 00:13:33.359234 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-68cf44c8b8-zkpbt" Mar 21 00:13:34 crc kubenswrapper[5117]: I0321 00:13:34.093139 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-747b44746d-7zctx" Mar 21 00:13:34 crc kubenswrapper[5117]: I0321 00:13:34.140678 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:13:35 crc kubenswrapper[5117]: E0321 00:13:35.150059 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 00:13:35 crc kubenswrapper[5117]: E0321 00:13:35.152243 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 00:13:35 crc kubenswrapper[5117]: E0321 00:13:35.153767 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 00:13:35 crc kubenswrapper[5117]: E0321 00:13:35.153813 5117 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" podUID="ee9b5fe1-7a59-49c5-b486-a6d83d56728c" containerName="kube-multus-additional-cni-plugins" probeResult="unknown" Mar 21 00:13:35 crc kubenswrapper[5117]: I0321 00:13:35.213681 5117 ???:1] "http: TLS handshake error from 192.168.126.11:49460: no serving certificate available for the kubelet" Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.327999 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.351039 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kube-api-access\") pod \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\" (UID: \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\") " Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.351211 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kubelet-dir\") pod \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\" (UID: \"b59d6290-e0a2-4a09-85c7-adafaa952dc9\") " Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.351513 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b59d6290-e0a2-4a09-85c7-adafaa952dc9" (UID: "b59d6290-e0a2-4a09-85c7-adafaa952dc9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.359732 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b59d6290-e0a2-4a09-85c7-adafaa952dc9" (UID: "b59d6290-e0a2-4a09-85c7-adafaa952dc9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.453359 5117 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.453392 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b59d6290-e0a2-4a09-85c7-adafaa952dc9-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.467225 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"b59d6290-e0a2-4a09-85c7-adafaa952dc9","Type":"ContainerDied","Data":"c8f7c97c82bd89d7641b19a2f6e3b59f18c883ce6b859ba55ad47b0d26355706"} Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.467563 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8f7c97c82bd89d7641b19a2f6e3b59f18c883ce6b859ba55ad47b0d26355706" Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.467675 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.963575 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-dx9lp"] Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.963889 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" podUID="a5a8eb05-2bff-483b-b42f-03b818f9a058" containerName="controller-manager" containerID="cri-o://55b428f845a56f759367d07dee79ae047819a764952b0ed1e3738688f5053052" gracePeriod=30 Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.987599 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk"] Mar 21 00:13:37 crc kubenswrapper[5117]: I0321 00:13:37.991269 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" podUID="a29e146a-be7b-49f8-b97b-31f4c976afb5" containerName="route-controller-manager" containerID="cri-o://d6b33be80794ace1b337dea7a890f0a2017717b2b887c92b65e9190537832921" gracePeriod=30 Mar 21 00:13:38 crc kubenswrapper[5117]: I0321 00:13:38.474876 5117 generic.go:358] "Generic (PLEG): container finished" podID="a5a8eb05-2bff-483b-b42f-03b818f9a058" containerID="55b428f845a56f759367d07dee79ae047819a764952b0ed1e3738688f5053052" exitCode=0 Mar 21 00:13:38 crc kubenswrapper[5117]: I0321 00:13:38.475003 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" event={"ID":"a5a8eb05-2bff-483b-b42f-03b818f9a058","Type":"ContainerDied","Data":"55b428f845a56f759367d07dee79ae047819a764952b0ed1e3738688f5053052"} Mar 21 00:13:38 crc kubenswrapper[5117]: I0321 00:13:38.478457 5117 generic.go:358] "Generic (PLEG): container finished" podID="a29e146a-be7b-49f8-b97b-31f4c976afb5" containerID="d6b33be80794ace1b337dea7a890f0a2017717b2b887c92b65e9190537832921" exitCode=0 Mar 21 00:13:38 crc kubenswrapper[5117]: I0321 00:13:38.478508 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" event={"ID":"a29e146a-be7b-49f8-b97b-31f4c976afb5","Type":"ContainerDied","Data":"d6b33be80794ace1b337dea7a890f0a2017717b2b887c92b65e9190537832921"} Mar 21 00:13:38 crc kubenswrapper[5117]: I0321 00:13:38.772440 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:38 crc kubenswrapper[5117]: I0321 00:13:38.778990 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-64d44f6ddf-dg2qw" Mar 21 00:13:41 crc kubenswrapper[5117]: I0321 00:13:41.561889 5117 patch_prober.go:28] interesting pod/controller-manager-65b6cccf98-dx9lp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 21 00:13:41 crc kubenswrapper[5117]: I0321 00:13:41.562323 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" podUID="a5a8eb05-2bff-483b-b42f-03b818f9a058" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.117162 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.166844 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db77b7694-njxld"] Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.167895 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b59d6290-e0a2-4a09-85c7-adafaa952dc9" containerName="pruner" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.167920 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59d6290-e0a2-4a09-85c7-adafaa952dc9" containerName="pruner" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.167956 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a29e146a-be7b-49f8-b97b-31f4c976afb5" containerName="route-controller-manager" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.167963 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29e146a-be7b-49f8-b97b-31f4c976afb5" containerName="route-controller-manager" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.167975 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece" containerName="pruner" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.167981 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece" containerName="pruner" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.168079 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="4f0d0bb0-dc06-47fe-9c0b-d8ee2025eece" containerName="pruner" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.168092 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="a29e146a-be7b-49f8-b97b-31f4c976afb5" containerName="route-controller-manager" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.168106 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="b59d6290-e0a2-4a09-85c7-adafaa952dc9" containerName="pruner" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.187423 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db77b7694-njxld"] Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.187640 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.231534 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a29e146a-be7b-49f8-b97b-31f4c976afb5-tmp\") pod \"a29e146a-be7b-49f8-b97b-31f4c976afb5\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.231624 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-config\") pod \"a29e146a-be7b-49f8-b97b-31f4c976afb5\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.231665 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29e146a-be7b-49f8-b97b-31f4c976afb5-serving-cert\") pod \"a29e146a-be7b-49f8-b97b-31f4c976afb5\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.231709 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-client-ca\") pod \"a29e146a-be7b-49f8-b97b-31f4c976afb5\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.231761 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pksg4\" (UniqueName: \"kubernetes.io/projected/a29e146a-be7b-49f8-b97b-31f4c976afb5-kube-api-access-pksg4\") pod \"a29e146a-be7b-49f8-b97b-31f4c976afb5\" (UID: \"a29e146a-be7b-49f8-b97b-31f4c976afb5\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.231972 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38b6e9-af77-49d5-bb5b-4341f66d3224-serving-cert\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.232010 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-client-ca\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.232042 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-config\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.232078 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ac38b6e9-af77-49d5-bb5b-4341f66d3224-tmp\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.232139 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcvsc\" (UniqueName: \"kubernetes.io/projected/ac38b6e9-af77-49d5-bb5b-4341f66d3224-kube-api-access-vcvsc\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.232203 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a29e146a-be7b-49f8-b97b-31f4c976afb5-tmp" (OuterVolumeSpecName: "tmp") pod "a29e146a-be7b-49f8-b97b-31f4c976afb5" (UID: "a29e146a-be7b-49f8-b97b-31f4c976afb5"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.233059 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-config" (OuterVolumeSpecName: "config") pod "a29e146a-be7b-49f8-b97b-31f4c976afb5" (UID: "a29e146a-be7b-49f8-b97b-31f4c976afb5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.233422 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-client-ca" (OuterVolumeSpecName: "client-ca") pod "a29e146a-be7b-49f8-b97b-31f4c976afb5" (UID: "a29e146a-be7b-49f8-b97b-31f4c976afb5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.242029 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a29e146a-be7b-49f8-b97b-31f4c976afb5-kube-api-access-pksg4" (OuterVolumeSpecName: "kube-api-access-pksg4") pod "a29e146a-be7b-49f8-b97b-31f4c976afb5" (UID: "a29e146a-be7b-49f8-b97b-31f4c976afb5"). InnerVolumeSpecName "kube-api-access-pksg4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.242973 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29e146a-be7b-49f8-b97b-31f4c976afb5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a29e146a-be7b-49f8-b97b-31f4c976afb5" (UID: "a29e146a-be7b-49f8-b97b-31f4c976afb5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.318274 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.333929 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vcvsc\" (UniqueName: \"kubernetes.io/projected/ac38b6e9-af77-49d5-bb5b-4341f66d3224-kube-api-access-vcvsc\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334066 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38b6e9-af77-49d5-bb5b-4341f66d3224-serving-cert\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334100 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-client-ca\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334141 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-config\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334185 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ac38b6e9-af77-49d5-bb5b-4341f66d3224-tmp\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334285 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pksg4\" (UniqueName: \"kubernetes.io/projected/a29e146a-be7b-49f8-b97b-31f4c976afb5-kube-api-access-pksg4\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334307 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a29e146a-be7b-49f8-b97b-31f4c976afb5-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334320 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334331 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29e146a-be7b-49f8-b97b-31f4c976afb5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334346 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a29e146a-be7b-49f8-b97b-31f4c976afb5-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.334952 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ac38b6e9-af77-49d5-bb5b-4341f66d3224-tmp\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.339972 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-config\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.340922 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-client-ca\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.340959 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38b6e9-af77-49d5-bb5b-4341f66d3224-serving-cert\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.356377 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcvsc\" (UniqueName: \"kubernetes.io/projected/ac38b6e9-af77-49d5-bb5b-4341f66d3224-kube-api-access-vcvsc\") pod \"route-controller-manager-db77b7694-njxld\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.371207 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7b449fd8cc-9ckff"] Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.371924 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a5a8eb05-2bff-483b-b42f-03b818f9a058" containerName="controller-manager" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.371945 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a8eb05-2bff-483b-b42f-03b818f9a058" containerName="controller-manager" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.372055 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="a5a8eb05-2bff-483b-b42f-03b818f9a058" containerName="controller-manager" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.390366 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b449fd8cc-9ckff"] Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.390518 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.438575 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a5a8eb05-2bff-483b-b42f-03b818f9a058-tmp\") pod \"a5a8eb05-2bff-483b-b42f-03b818f9a058\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.438664 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-config\") pod \"a5a8eb05-2bff-483b-b42f-03b818f9a058\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.438698 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r878v\" (UniqueName: \"kubernetes.io/projected/a5a8eb05-2bff-483b-b42f-03b818f9a058-kube-api-access-r878v\") pod \"a5a8eb05-2bff-483b-b42f-03b818f9a058\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.438723 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-client-ca\") pod \"a5a8eb05-2bff-483b-b42f-03b818f9a058\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.438757 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5a8eb05-2bff-483b-b42f-03b818f9a058-serving-cert\") pod \"a5a8eb05-2bff-483b-b42f-03b818f9a058\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.438820 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-proxy-ca-bundles\") pod \"a5a8eb05-2bff-483b-b42f-03b818f9a058\" (UID: \"a5a8eb05-2bff-483b-b42f-03b818f9a058\") " Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.439073 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-serving-cert\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.439104 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8z4b\" (UniqueName: \"kubernetes.io/projected/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-kube-api-access-f8z4b\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.439126 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-client-ca\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.439146 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-tmp\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.439165 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-config\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.439205 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-proxy-ca-bundles\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.439634 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5a8eb05-2bff-483b-b42f-03b818f9a058-tmp" (OuterVolumeSpecName: "tmp") pod "a5a8eb05-2bff-483b-b42f-03b818f9a058" (UID: "a5a8eb05-2bff-483b-b42f-03b818f9a058"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.440389 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-config" (OuterVolumeSpecName: "config") pod "a5a8eb05-2bff-483b-b42f-03b818f9a058" (UID: "a5a8eb05-2bff-483b-b42f-03b818f9a058"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.440915 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-client-ca" (OuterVolumeSpecName: "client-ca") pod "a5a8eb05-2bff-483b-b42f-03b818f9a058" (UID: "a5a8eb05-2bff-483b-b42f-03b818f9a058"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.441800 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a5a8eb05-2bff-483b-b42f-03b818f9a058" (UID: "a5a8eb05-2bff-483b-b42f-03b818f9a058"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.447889 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a8eb05-2bff-483b-b42f-03b818f9a058-kube-api-access-r878v" (OuterVolumeSpecName: "kube-api-access-r878v") pod "a5a8eb05-2bff-483b-b42f-03b818f9a058" (UID: "a5a8eb05-2bff-483b-b42f-03b818f9a058"). InnerVolumeSpecName "kube-api-access-r878v". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.448964 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a8eb05-2bff-483b-b42f-03b818f9a058-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a5a8eb05-2bff-483b-b42f-03b818f9a058" (UID: "a5a8eb05-2bff-483b-b42f-03b818f9a058"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.502631 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" event={"ID":"a5a8eb05-2bff-483b-b42f-03b818f9a058","Type":"ContainerDied","Data":"df46ff372feb0a56a5e8d86f9ba8582bbe49ad744382c4df90099fdf1892bda8"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.502714 5117 scope.go:117] "RemoveContainer" containerID="55b428f845a56f759367d07dee79ae047819a764952b0ed1e3738688f5053052" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.502978 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-dx9lp" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.507440 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" event={"ID":"a29e146a-be7b-49f8-b97b-31f4c976afb5","Type":"ContainerDied","Data":"34c3f5ed25dce25348c1c06732daffb81ae4d5ab489fe4b58efdeaf4a4996683"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.507486 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.521541 5117 generic.go:358] "Generic (PLEG): container finished" podID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerID="88171e4aebd38e5bd7a053713b0629563b8fc4583ca64e03e8db7d6616476eed" exitCode=0 Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.521708 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz8n5" event={"ID":"55a3135d-1190-4674-9ceb-a8d36fc19e0b","Type":"ContainerDied","Data":"88171e4aebd38e5bd7a053713b0629563b8fc4583ca64e03e8db7d6616476eed"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.523017 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.529294 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mz5qp" event={"ID":"30a52c1e-3bfa-4789-a6ab-d851d6139934","Type":"ContainerStarted","Data":"3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.536300 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" event={"ID":"f85237c3-327f-4a9e-8496-00c9087e5aa7","Type":"ContainerStarted","Data":"c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.539111 5117 generic.go:358] "Generic (PLEG): container finished" podID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerID="caa984f8fd4f103859bf67dd88c6c152ef152721e4f0a0ba01b514c9adda7978" exitCode=0 Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.539260 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lgn8x" event={"ID":"1eac1109-f55b-4a75-9d83-ad0aba830baa","Type":"ContainerDied","Data":"caa984f8fd4f103859bf67dd88c6c152ef152721e4f0a0ba01b514c9adda7978"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.546691 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.548143 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-serving-cert\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.548220 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f8z4b\" (UniqueName: \"kubernetes.io/projected/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-kube-api-access-f8z4b\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.548280 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-client-ca\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.548306 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-tmp\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.548326 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-config\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.548402 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-proxy-ca-bundles\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.550232 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.550474 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.550956 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.557319 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqkbn" event={"ID":"a106308f-08b0-4f04-a431-8aa7042e4fbb","Type":"ContainerStarted","Data":"b254996c08a74ae7d0a586b1ce8b34673c9b8de5d9bc0a68c55b18f1da5432e8"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.577767 5117 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.578561 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-tmp\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.579588 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a5a8eb05-2bff-483b-b42f-03b818f9a058-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.579695 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-config\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.579733 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.579778 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-r878v\" (UniqueName: \"kubernetes.io/projected/a5a8eb05-2bff-483b-b42f-03b818f9a058-kube-api-access-r878v\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.579898 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5a8eb05-2bff-483b-b42f-03b818f9a058-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.579982 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-client-ca\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.580212 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5a8eb05-2bff-483b-b42f-03b818f9a058-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.587466 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-serving-cert\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.599523 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-proxy-ca-bundles\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.602740 5117 scope.go:117] "RemoveContainer" containerID="d6b33be80794ace1b337dea7a890f0a2017717b2b887c92b65e9190537832921" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.612079 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8z4b\" (UniqueName: \"kubernetes.io/projected/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-kube-api-access-f8z4b\") pod \"controller-manager-7b449fd8cc-9ckff\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.625873 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz57f" event={"ID":"c15a6f5e-ffea-444b-98d1-3922aeb22128","Type":"ContainerStarted","Data":"61db64811d8fa48f4dea7bc5e6b24139fa0f2fa85efb6e7f6637840a6d8953e9"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.646061 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=49.646040302 podStartE2EDuration="49.646040302s" podCreationTimestamp="2026-03-21 00:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:42.632405522 +0000 UTC m=+135.925692704" watchObservedRunningTime="2026-03-21 00:13:42.646040302 +0000 UTC m=+135.939327474" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.697644 5117 generic.go:358] "Generic (PLEG): container finished" podID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerID="e74f027a89bba3cb5cc561e52ac36add0f5d4840e394010297464154ecbf26fb" exitCode=0 Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.697782 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn6n6" event={"ID":"0c19bd60-8272-4e5e-81d3-10660df3c5ed","Type":"ContainerDied","Data":"e74f027a89bba3cb5cc561e52ac36add0f5d4840e394010297464154ecbf26fb"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.705397 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-dx9lp"] Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.715792 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h696r" event={"ID":"4b71d84d-21ab-478c-9d5e-4ca7c422b358","Type":"ContainerStarted","Data":"06ab4a88c8c400805a4de52ba976f0dd453eda222a4aeee879e746411f2cd7db"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.723392 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2v8t" event={"ID":"0c7aa7ad-faf4-41a3-82a1-c064c8109629","Type":"ContainerStarted","Data":"81b5faf2fb3f54852f02594fe0d6888d914a0752b0d59ab5476442da7de5e373"} Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.731807 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-dx9lp"] Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.770305 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk"] Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.770888 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-l6ptk"] Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.806767 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:42 crc kubenswrapper[5117]: I0321 00:13:42.826975 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" podStartSLOduration=114.82695217 podStartE2EDuration="1m54.82695217s" podCreationTimestamp="2026-03-21 00:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:42.826378122 +0000 UTC m=+136.119665304" watchObservedRunningTime="2026-03-21 00:13:42.82695217 +0000 UTC m=+136.120239342" Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.069159 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b449fd8cc-9ckff"] Mar 21 00:13:43 crc kubenswrapper[5117]: W0321 00:13:43.077083 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0c28656_8b42_4beb_be05_b5c0ee7cdca4.slice/crio-7648e6dbc8de7fada85d45f9f1cedd8bd90573bab102ee64ec07c10789567066 WatchSource:0}: Error finding container 7648e6dbc8de7fada85d45f9f1cedd8bd90573bab102ee64ec07c10789567066: Status 404 returned error can't find the container with id 7648e6dbc8de7fada85d45f9f1cedd8bd90573bab102ee64ec07c10789567066 Mar 21 00:13:43 crc kubenswrapper[5117]: W0321 00:13:43.172641 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac38b6e9_af77_49d5_bb5b_4341f66d3224.slice/crio-09f3f77e2be372aa234c97865bc71ceca91acb856e31f9dfd514f83b862de9bd WatchSource:0}: Error finding container 09f3f77e2be372aa234c97865bc71ceca91acb856e31f9dfd514f83b862de9bd: Status 404 returned error can't find the container with id 09f3f77e2be372aa234c97865bc71ceca91acb856e31f9dfd514f83b862de9bd Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.175193 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db77b7694-njxld"] Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.591115 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a29e146a-be7b-49f8-b97b-31f4c976afb5" path="/var/lib/kubelet/pods/a29e146a-be7b-49f8-b97b-31f4c976afb5/volumes" Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.591856 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5a8eb05-2bff-483b-b42f-03b818f9a058" path="/var/lib/kubelet/pods/a5a8eb05-2bff-483b-b42f-03b818f9a058/volumes" Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.730668 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" event={"ID":"ac38b6e9-af77-49d5-bb5b-4341f66d3224","Type":"ContainerStarted","Data":"09f3f77e2be372aa234c97865bc71ceca91acb856e31f9dfd514f83b862de9bd"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.732754 5117 generic.go:358] "Generic (PLEG): container finished" podID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerID="3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8" exitCode=0 Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.732870 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mz5qp" event={"ID":"30a52c1e-3bfa-4789-a6ab-d851d6139934","Type":"ContainerDied","Data":"3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.737308 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lgn8x" event={"ID":"1eac1109-f55b-4a75-9d83-ad0aba830baa","Type":"ContainerStarted","Data":"fb1fd2789f4cb7019f8780db05055d4abaaf9b385a76e910e3aaba35a9cd48df"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.738908 5117 generic.go:358] "Generic (PLEG): container finished" podID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerID="b254996c08a74ae7d0a586b1ce8b34673c9b8de5d9bc0a68c55b18f1da5432e8" exitCode=0 Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.739001 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqkbn" event={"ID":"a106308f-08b0-4f04-a431-8aa7042e4fbb","Type":"ContainerDied","Data":"b254996c08a74ae7d0a586b1ce8b34673c9b8de5d9bc0a68c55b18f1da5432e8"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.741133 5117 generic.go:358] "Generic (PLEG): container finished" podID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerID="61db64811d8fa48f4dea7bc5e6b24139fa0f2fa85efb6e7f6637840a6d8953e9" exitCode=0 Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.741223 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz57f" event={"ID":"c15a6f5e-ffea-444b-98d1-3922aeb22128","Type":"ContainerDied","Data":"61db64811d8fa48f4dea7bc5e6b24139fa0f2fa85efb6e7f6637840a6d8953e9"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.743008 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" event={"ID":"c0c28656-8b42-4beb-be05-b5c0ee7cdca4","Type":"ContainerStarted","Data":"7648e6dbc8de7fada85d45f9f1cedd8bd90573bab102ee64ec07c10789567066"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.746892 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn6n6" event={"ID":"0c19bd60-8272-4e5e-81d3-10660df3c5ed","Type":"ContainerStarted","Data":"7b584406c0cc2175e43e4ba73514501a4feb96a61c59a40c9fb6a04b8c19b657"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.749195 5117 generic.go:358] "Generic (PLEG): container finished" podID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerID="06ab4a88c8c400805a4de52ba976f0dd453eda222a4aeee879e746411f2cd7db" exitCode=0 Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.749290 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h696r" event={"ID":"4b71d84d-21ab-478c-9d5e-4ca7c422b358","Type":"ContainerDied","Data":"06ab4a88c8c400805a4de52ba976f0dd453eda222a4aeee879e746411f2cd7db"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.749345 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h696r" event={"ID":"4b71d84d-21ab-478c-9d5e-4ca7c422b358","Type":"ContainerStarted","Data":"e284537a540e4939d60b3a4b2835a3865c343963ee887e615ece42eac4eaa412"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.755731 5117 generic.go:358] "Generic (PLEG): container finished" podID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerID="81b5faf2fb3f54852f02594fe0d6888d914a0752b0d59ab5476442da7de5e373" exitCode=0 Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.755866 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2v8t" event={"ID":"0c7aa7ad-faf4-41a3-82a1-c064c8109629","Type":"ContainerDied","Data":"81b5faf2fb3f54852f02594fe0d6888d914a0752b0d59ab5476442da7de5e373"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.764540 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz8n5" event={"ID":"55a3135d-1190-4674-9ceb-a8d36fc19e0b","Type":"ContainerStarted","Data":"84d49c227c93caca5e26dd594ae063ee8daf62ad0455825e44461224e4598890"} Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.816057 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h696r" podStartSLOduration=5.238151045 podStartE2EDuration="26.816036791s" podCreationTimestamp="2026-03-21 00:13:17 +0000 UTC" firstStartedPulling="2026-03-21 00:13:20.218803123 +0000 UTC m=+113.512090295" lastFinishedPulling="2026-03-21 00:13:41.796688879 +0000 UTC m=+135.089976041" observedRunningTime="2026-03-21 00:13:43.81084098 +0000 UTC m=+137.104128162" watchObservedRunningTime="2026-03-21 00:13:43.816036791 +0000 UTC m=+137.109323953" Mar 21 00:13:43 crc kubenswrapper[5117]: I0321 00:13:43.865662 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bn6n6" podStartSLOduration=4.320962916 podStartE2EDuration="24.86564443s" podCreationTimestamp="2026-03-21 00:13:19 +0000 UTC" firstStartedPulling="2026-03-21 00:13:21.287443797 +0000 UTC m=+114.580730969" lastFinishedPulling="2026-03-21 00:13:41.832125311 +0000 UTC m=+135.125412483" observedRunningTime="2026-03-21 00:13:43.8409763 +0000 UTC m=+137.134263472" watchObservedRunningTime="2026-03-21 00:13:43.86564443 +0000 UTC m=+137.158931592" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.772895 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqkbn" event={"ID":"a106308f-08b0-4f04-a431-8aa7042e4fbb","Type":"ContainerStarted","Data":"250f928cbfbe573d75421660ff381149b350fb2932ca146f21b1bf2224c43eef"} Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.775757 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz57f" event={"ID":"c15a6f5e-ffea-444b-98d1-3922aeb22128","Type":"ContainerStarted","Data":"8bfa5e41be23006122c4f02a480d001bccc38bb00d52d38a26ee119322b434b9"} Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.778429 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" event={"ID":"c0c28656-8b42-4beb-be05-b5c0ee7cdca4","Type":"ContainerStarted","Data":"f58bd6065bdfd1b3bf07376d7d094f3ad1ead1ad9ba24fc20d100db0e091587d"} Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.779154 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.781182 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2v8t" event={"ID":"0c7aa7ad-faf4-41a3-82a1-c064c8109629","Type":"ContainerStarted","Data":"c60254e7c5eeb03728b5c9bf521ddf0fc6be4adb1d91a090ddec51ceac589d24"} Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.784178 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" event={"ID":"ac38b6e9-af77-49d5-bb5b-4341f66d3224","Type":"ContainerStarted","Data":"874da865045fae9905bc80bfe3ae82b9e275936a1f84be89734dfa4c6794e6ae"} Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.784658 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.786679 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mz5qp" event={"ID":"30a52c1e-3bfa-4789-a6ab-d851d6139934","Type":"ContainerStarted","Data":"0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe"} Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.792240 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.795556 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rz8n5" podStartSLOduration=5.25714186 podStartE2EDuration="25.795536147s" podCreationTimestamp="2026-03-21 00:13:19 +0000 UTC" firstStartedPulling="2026-03-21 00:13:21.30849852 +0000 UTC m=+114.601785692" lastFinishedPulling="2026-03-21 00:13:41.846892807 +0000 UTC m=+135.140179979" observedRunningTime="2026-03-21 00:13:44.068238875 +0000 UTC m=+137.361526067" watchObservedRunningTime="2026-03-21 00:13:44.795536147 +0000 UTC m=+138.088823319" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.797208 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pqkbn" podStartSLOduration=5.155537906 podStartE2EDuration="27.797201168s" podCreationTimestamp="2026-03-21 00:13:17 +0000 UTC" firstStartedPulling="2026-03-21 00:13:19.155021357 +0000 UTC m=+112.448308529" lastFinishedPulling="2026-03-21 00:13:41.796684619 +0000 UTC m=+135.089971791" observedRunningTime="2026-03-21 00:13:44.793534365 +0000 UTC m=+138.086821537" watchObservedRunningTime="2026-03-21 00:13:44.797201168 +0000 UTC m=+138.090488340" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.824201 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" podStartSLOduration=7.824180609 podStartE2EDuration="7.824180609s" podCreationTimestamp="2026-03-21 00:13:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:44.82094507 +0000 UTC m=+138.114232232" watchObservedRunningTime="2026-03-21 00:13:44.824180609 +0000 UTC m=+138.117467781" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.859488 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d2v8t" podStartSLOduration=5.284202798 podStartE2EDuration="24.859465968s" podCreationTimestamp="2026-03-21 00:13:20 +0000 UTC" firstStartedPulling="2026-03-21 00:13:22.322792404 +0000 UTC m=+115.616079566" lastFinishedPulling="2026-03-21 00:13:41.898055574 +0000 UTC m=+135.191342736" observedRunningTime="2026-03-21 00:13:44.858542549 +0000 UTC m=+138.151829731" watchObservedRunningTime="2026-03-21 00:13:44.859465968 +0000 UTC m=+138.152753140" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.880933 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.888736 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lgn8x" podStartSLOduration=5.187132786 podStartE2EDuration="27.888713999s" podCreationTimestamp="2026-03-21 00:13:17 +0000 UTC" firstStartedPulling="2026-03-21 00:13:19.134903553 +0000 UTC m=+112.428190725" lastFinishedPulling="2026-03-21 00:13:41.836484736 +0000 UTC m=+135.129771938" observedRunningTime="2026-03-21 00:13:44.888201704 +0000 UTC m=+138.181488886" watchObservedRunningTime="2026-03-21 00:13:44.888713999 +0000 UTC m=+138.182001171" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.926671 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qz57f" podStartSLOduration=5.401372291 podStartE2EDuration="24.926650749s" podCreationTimestamp="2026-03-21 00:13:20 +0000 UTC" firstStartedPulling="2026-03-21 00:13:22.317660287 +0000 UTC m=+115.610947459" lastFinishedPulling="2026-03-21 00:13:41.842938745 +0000 UTC m=+135.136225917" observedRunningTime="2026-03-21 00:13:44.924996637 +0000 UTC m=+138.218283809" watchObservedRunningTime="2026-03-21 00:13:44.926650749 +0000 UTC m=+138.219937921" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.947731 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" podStartSLOduration=6.947709888 podStartE2EDuration="6.947709888s" podCreationTimestamp="2026-03-21 00:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:13:44.946535972 +0000 UTC m=+138.239823154" watchObservedRunningTime="2026-03-21 00:13:44.947709888 +0000 UTC m=+138.240997060" Mar 21 00:13:44 crc kubenswrapper[5117]: I0321 00:13:44.978701 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mz5qp" podStartSLOduration=6.368311202 podStartE2EDuration="27.978680802s" podCreationTimestamp="2026-03-21 00:13:17 +0000 UTC" firstStartedPulling="2026-03-21 00:13:20.227166088 +0000 UTC m=+113.520453260" lastFinishedPulling="2026-03-21 00:13:41.837535688 +0000 UTC m=+135.130822860" observedRunningTime="2026-03-21 00:13:44.976376942 +0000 UTC m=+138.269664114" watchObservedRunningTime="2026-03-21 00:13:44.978680802 +0000 UTC m=+138.271967984" Mar 21 00:13:45 crc kubenswrapper[5117]: E0321 00:13:45.150783 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 00:13:45 crc kubenswrapper[5117]: E0321 00:13:45.152310 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 00:13:45 crc kubenswrapper[5117]: E0321 00:13:45.153572 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 21 00:13:45 crc kubenswrapper[5117]: E0321 00:13:45.153660 5117 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" podUID="ee9b5fe1-7a59-49c5-b486-a6d83d56728c" containerName="kube-multus-additional-cni-plugins" probeResult="unknown" Mar 21 00:13:46 crc kubenswrapper[5117]: I0321 00:13:46.907963 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-9zrjr" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.470865 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.470940 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.697525 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.697704 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.738444 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.755711 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.884583 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.884782 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.930839 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:47 crc kubenswrapper[5117]: I0321 00:13:47.953298 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-fhkjl" Mar 21 00:13:48 crc kubenswrapper[5117]: I0321 00:13:48.125121 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:48 crc kubenswrapper[5117]: I0321 00:13:48.125176 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:48 crc kubenswrapper[5117]: I0321 00:13:48.162468 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:48 crc kubenswrapper[5117]: I0321 00:13:48.812240 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-jjwjr_ee9b5fe1-7a59-49c5-b486-a6d83d56728c/kube-multus-additional-cni-plugins/0.log" Mar 21 00:13:48 crc kubenswrapper[5117]: I0321 00:13:48.812763 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee9b5fe1-7a59-49c5-b486-a6d83d56728c" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" exitCode=137 Mar 21 00:13:48 crc kubenswrapper[5117]: I0321 00:13:48.812863 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" event={"ID":"ee9b5fe1-7a59-49c5-b486-a6d83d56728c","Type":"ContainerDied","Data":"1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182"} Mar 21 00:13:48 crc kubenswrapper[5117]: I0321 00:13:48.873925 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:48 crc kubenswrapper[5117]: I0321 00:13:48.880109 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-jjwjr_ee9b5fe1-7a59-49c5-b486-a6d83d56728c/kube-multus-additional-cni-plugins/0.log" Mar 21 00:13:48 crc kubenswrapper[5117]: I0321 00:13:48.880206 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.009745 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-ready\") pod \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.009985 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-cni-sysctl-allowlist\") pod \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.010030 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dchfn\" (UniqueName: \"kubernetes.io/projected/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-kube-api-access-dchfn\") pod \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.010046 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-tuning-conf-dir\") pod \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\" (UID: \"ee9b5fe1-7a59-49c5-b486-a6d83d56728c\") " Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.010717 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-ready" (OuterVolumeSpecName: "ready") pod "ee9b5fe1-7a59-49c5-b486-a6d83d56728c" (UID: "ee9b5fe1-7a59-49c5-b486-a6d83d56728c"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.011070 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "ee9b5fe1-7a59-49c5-b486-a6d83d56728c" (UID: "ee9b5fe1-7a59-49c5-b486-a6d83d56728c"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.011147 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "ee9b5fe1-7a59-49c5-b486-a6d83d56728c" (UID: "ee9b5fe1-7a59-49c5-b486-a6d83d56728c"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.018396 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-kube-api-access-dchfn" (OuterVolumeSpecName: "kube-api-access-dchfn") pod "ee9b5fe1-7a59-49c5-b486-a6d83d56728c" (UID: "ee9b5fe1-7a59-49c5-b486-a6d83d56728c"). InnerVolumeSpecName "kube-api-access-dchfn". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.111350 5117 reconciler_common.go:299] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-ready\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.111395 5117 reconciler_common.go:299] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.111407 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dchfn\" (UniqueName: \"kubernetes.io/projected/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-kube-api-access-dchfn\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.111418 5117 reconciler_common.go:299] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ee9b5fe1-7a59-49c5-b486-a6d83d56728c-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.711639 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.711728 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.760344 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.820452 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-jjwjr_ee9b5fe1-7a59-49c5-b486-a6d83d56728c/kube-multus-additional-cni-plugins/0.log" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.821224 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" event={"ID":"ee9b5fe1-7a59-49c5-b486-a6d83d56728c","Type":"ContainerDied","Data":"e106ab1dce81c279867542bdfef7aad624495b6b44728187c9d85911e171c5e4"} Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.821311 5117 scope.go:117] "RemoveContainer" containerID="1cc056e27e4cc717480d0b3825e474cce3c01b8c469a8e9a68577b193af73182" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.821504 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-jjwjr" Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.853354 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-jjwjr"] Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.855340 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-jjwjr"] Mar 21 00:13:49 crc kubenswrapper[5117]: I0321 00:13:49.875794 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:13:50 crc kubenswrapper[5117]: I0321 00:13:50.081958 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:50 crc kubenswrapper[5117]: I0321 00:13:50.082009 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:50 crc kubenswrapper[5117]: I0321 00:13:50.127859 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:50 crc kubenswrapper[5117]: I0321 00:13:50.695026 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:50 crc kubenswrapper[5117]: I0321 00:13:50.695146 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:13:50 crc kubenswrapper[5117]: I0321 00:13:50.987531 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:51 crc kubenswrapper[5117]: I0321 00:13:51.080279 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:51 crc kubenswrapper[5117]: I0321 00:13:51.080339 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:13:51 crc kubenswrapper[5117]: I0321 00:13:51.357425 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h696r"] Mar 21 00:13:51 crc kubenswrapper[5117]: I0321 00:13:51.586339 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9b5fe1-7a59-49c5-b486-a6d83d56728c" path="/var/lib/kubelet/pods/ee9b5fe1-7a59-49c5-b486-a6d83d56728c/volumes" Mar 21 00:13:51 crc kubenswrapper[5117]: I0321 00:13:51.736947 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d2v8t" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerName="registry-server" probeResult="failure" output=< Mar 21 00:13:51 crc kubenswrapper[5117]: timeout: failed to connect service ":50051" within 1s Mar 21 00:13:51 crc kubenswrapper[5117]: > Mar 21 00:13:51 crc kubenswrapper[5117]: I0321 00:13:51.834512 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h696r" podUID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerName="registry-server" containerID="cri-o://e284537a540e4939d60b3a4b2835a3865c343963ee887e615ece42eac4eaa412" gracePeriod=2 Mar 21 00:13:52 crc kubenswrapper[5117]: I0321 00:13:52.117606 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qz57f" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerName="registry-server" probeResult="failure" output=< Mar 21 00:13:52 crc kubenswrapper[5117]: timeout: failed to connect service ":50051" within 1s Mar 21 00:13:52 crc kubenswrapper[5117]: > Mar 21 00:13:53 crc kubenswrapper[5117]: I0321 00:13:53.754739 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz8n5"] Mar 21 00:13:53 crc kubenswrapper[5117]: I0321 00:13:53.755416 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rz8n5" podUID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerName="registry-server" containerID="cri-o://84d49c227c93caca5e26dd594ae063ee8daf62ad0455825e44461224e4598890" gracePeriod=2 Mar 21 00:13:53 crc kubenswrapper[5117]: I0321 00:13:53.851874 5117 generic.go:358] "Generic (PLEG): container finished" podID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerID="e284537a540e4939d60b3a4b2835a3865c343963ee887e615ece42eac4eaa412" exitCode=0 Mar 21 00:13:53 crc kubenswrapper[5117]: I0321 00:13:53.851941 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h696r" event={"ID":"4b71d84d-21ab-478c-9d5e-4ca7c422b358","Type":"ContainerDied","Data":"e284537a540e4939d60b3a4b2835a3865c343963ee887e615ece42eac4eaa412"} Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.208110 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.288432 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-catalog-content\") pod \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.288855 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-utilities\") pod \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.289153 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wndv\" (UniqueName: \"kubernetes.io/projected/4b71d84d-21ab-478c-9d5e-4ca7c422b358-kube-api-access-4wndv\") pod \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\" (UID: \"4b71d84d-21ab-478c-9d5e-4ca7c422b358\") " Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.289974 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-utilities" (OuterVolumeSpecName: "utilities") pod "4b71d84d-21ab-478c-9d5e-4ca7c422b358" (UID: "4b71d84d-21ab-478c-9d5e-4ca7c422b358"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.295163 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b71d84d-21ab-478c-9d5e-4ca7c422b358-kube-api-access-4wndv" (OuterVolumeSpecName: "kube-api-access-4wndv") pod "4b71d84d-21ab-478c-9d5e-4ca7c422b358" (UID: "4b71d84d-21ab-478c-9d5e-4ca7c422b358"). InnerVolumeSpecName "kube-api-access-4wndv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.359477 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b71d84d-21ab-478c-9d5e-4ca7c422b358" (UID: "4b71d84d-21ab-478c-9d5e-4ca7c422b358"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.390854 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4wndv\" (UniqueName: \"kubernetes.io/projected/4b71d84d-21ab-478c-9d5e-4ca7c422b358-kube-api-access-4wndv\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.390896 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.390913 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b71d84d-21ab-478c-9d5e-4ca7c422b358-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.408788 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.861124 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h696r" event={"ID":"4b71d84d-21ab-478c-9d5e-4ca7c422b358","Type":"ContainerDied","Data":"07062a0dfcfa62fc9b1b51cc65647063fd403489ca12603b798dbdc17e77cc28"} Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.861982 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h696r" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.863421 5117 scope.go:117] "RemoveContainer" containerID="e284537a540e4939d60b3a4b2835a3865c343963ee887e615ece42eac4eaa412" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.871194 5117 generic.go:358] "Generic (PLEG): container finished" podID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerID="84d49c227c93caca5e26dd594ae063ee8daf62ad0455825e44461224e4598890" exitCode=0 Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.871413 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz8n5" event={"ID":"55a3135d-1190-4674-9ceb-a8d36fc19e0b","Type":"ContainerDied","Data":"84d49c227c93caca5e26dd594ae063ee8daf62ad0455825e44461224e4598890"} Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.904214 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h696r"] Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.906155 5117 scope.go:117] "RemoveContainer" containerID="06ab4a88c8c400805a4de52ba976f0dd453eda222a4aeee879e746411f2cd7db" Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.910325 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h696r"] Mar 21 00:13:54 crc kubenswrapper[5117]: I0321 00:13:54.924304 5117 scope.go:117] "RemoveContainer" containerID="feaab3061bfc5e8f9424e93af4c9a1af21b957460251a8a8d64aba5c6376d387" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.124640 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.204697 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-catalog-content\") pod \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.213422 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-utilities\") pod \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.213542 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjmm9\" (UniqueName: \"kubernetes.io/projected/55a3135d-1190-4674-9ceb-a8d36fc19e0b-kube-api-access-cjmm9\") pod \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\" (UID: \"55a3135d-1190-4674-9ceb-a8d36fc19e0b\") " Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.215311 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-utilities" (OuterVolumeSpecName: "utilities") pod "55a3135d-1190-4674-9ceb-a8d36fc19e0b" (UID: "55a3135d-1190-4674-9ceb-a8d36fc19e0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.221363 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55a3135d-1190-4674-9ceb-a8d36fc19e0b-kube-api-access-cjmm9" (OuterVolumeSpecName: "kube-api-access-cjmm9") pod "55a3135d-1190-4674-9ceb-a8d36fc19e0b" (UID: "55a3135d-1190-4674-9ceb-a8d36fc19e0b"). InnerVolumeSpecName "kube-api-access-cjmm9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.226312 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55a3135d-1190-4674-9ceb-a8d36fc19e0b" (UID: "55a3135d-1190-4674-9ceb-a8d36fc19e0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.315309 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.315743 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a3135d-1190-4674-9ceb-a8d36fc19e0b-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.315755 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-cjmm9\" (UniqueName: \"kubernetes.io/projected/55a3135d-1190-4674-9ceb-a8d36fc19e0b-kube-api-access-cjmm9\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.587689 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" path="/var/lib/kubelet/pods/4b71d84d-21ab-478c-9d5e-4ca7c422b358/volumes" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.714624 5117 ???:1] "http: TLS handshake error from 192.168.126.11:42482: no serving certificate available for the kubelet" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.880415 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz8n5" event={"ID":"55a3135d-1190-4674-9ceb-a8d36fc19e0b","Type":"ContainerDied","Data":"d8f39500c5cda00616dd99e61e41be69293decf722d3e222fb4afbc7e8b30688"} Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.880507 5117 scope.go:117] "RemoveContainer" containerID="84d49c227c93caca5e26dd594ae063ee8daf62ad0455825e44461224e4598890" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.880503 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz8n5" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.902774 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz8n5"] Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.904041 5117 scope.go:117] "RemoveContainer" containerID="88171e4aebd38e5bd7a053713b0629563b8fc4583ca64e03e8db7d6616476eed" Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.907109 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz8n5"] Mar 21 00:13:55 crc kubenswrapper[5117]: I0321 00:13:55.923905 5117 scope.go:117] "RemoveContainer" containerID="14d9cae71ca5afd9df48f906ff026fd51a53648a761b927c5800ea835e85ba36" Mar 21 00:13:57 crc kubenswrapper[5117]: I0321 00:13:57.587758 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" path="/var/lib/kubelet/pods/55a3135d-1190-4674-9ceb-a8d36fc19e0b/volumes" Mar 21 00:13:57 crc kubenswrapper[5117]: I0321 00:13:57.794221 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mtrnq"] Mar 21 00:13:57 crc kubenswrapper[5117]: I0321 00:13:57.871940 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.006070 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b449fd8cc-9ckff"] Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.006420 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" podUID="c0c28656-8b42-4beb-be05-b5c0ee7cdca4" containerName="controller-manager" containerID="cri-o://f58bd6065bdfd1b3bf07376d7d094f3ad1ead1ad9ba24fc20d100db0e091587d" gracePeriod=30 Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.017968 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db77b7694-njxld"] Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.018313 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" podUID="ac38b6e9-af77-49d5-bb5b-4341f66d3224" containerName="route-controller-manager" containerID="cri-o://874da865045fae9905bc80bfe3ae82b9e275936a1f84be89734dfa4c6794e6ae" gracePeriod=30 Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.856733 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.861506 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.914537 5117 generic.go:358] "Generic (PLEG): container finished" podID="ac38b6e9-af77-49d5-bb5b-4341f66d3224" containerID="874da865045fae9905bc80bfe3ae82b9e275936a1f84be89734dfa4c6794e6ae" exitCode=0 Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.914755 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" event={"ID":"ac38b6e9-af77-49d5-bb5b-4341f66d3224","Type":"ContainerDied","Data":"874da865045fae9905bc80bfe3ae82b9e275936a1f84be89734dfa4c6794e6ae"} Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.916790 5117 generic.go:358] "Generic (PLEG): container finished" podID="c0c28656-8b42-4beb-be05-b5c0ee7cdca4" containerID="f58bd6065bdfd1b3bf07376d7d094f3ad1ead1ad9ba24fc20d100db0e091587d" exitCode=0 Mar 21 00:13:58 crc kubenswrapper[5117]: I0321 00:13:58.916830 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" event={"ID":"c0c28656-8b42-4beb-be05-b5c0ee7cdca4","Type":"ContainerDied","Data":"f58bd6065bdfd1b3bf07376d7d094f3ad1ead1ad9ba24fc20d100db0e091587d"} Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.060908 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.090439 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091073 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerName="extract-content" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091097 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerName="extract-content" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091123 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerName="registry-server" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091132 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerName="registry-server" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091154 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerName="extract-utilities" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091162 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerName="extract-utilities" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091173 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee9b5fe1-7a59-49c5-b486-a6d83d56728c" containerName="kube-multus-additional-cni-plugins" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091179 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9b5fe1-7a59-49c5-b486-a6d83d56728c" containerName="kube-multus-additional-cni-plugins" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091189 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerName="extract-content" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091194 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerName="extract-content" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091200 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ac38b6e9-af77-49d5-bb5b-4341f66d3224" containerName="route-controller-manager" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091205 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac38b6e9-af77-49d5-bb5b-4341f66d3224" containerName="route-controller-manager" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091213 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerName="registry-server" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091219 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerName="registry-server" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091232 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerName="extract-utilities" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091238 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerName="extract-utilities" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091369 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee9b5fe1-7a59-49c5-b486-a6d83d56728c" containerName="kube-multus-additional-cni-plugins" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091381 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="4b71d84d-21ab-478c-9d5e-4ca7c422b358" containerName="registry-server" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091390 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ac38b6e9-af77-49d5-bb5b-4341f66d3224" containerName="route-controller-manager" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.091399 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="55a3135d-1190-4674-9ceb-a8d36fc19e0b" containerName="registry-server" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.094588 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.108499 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.176418 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-client-ca\") pod \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.176463 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-config\") pod \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.176577 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcvsc\" (UniqueName: \"kubernetes.io/projected/ac38b6e9-af77-49d5-bb5b-4341f66d3224-kube-api-access-vcvsc\") pod \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.176662 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ac38b6e9-af77-49d5-bb5b-4341f66d3224-tmp\") pod \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.176707 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38b6e9-af77-49d5-bb5b-4341f66d3224-serving-cert\") pod \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\" (UID: \"ac38b6e9-af77-49d5-bb5b-4341f66d3224\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.176914 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgg6c\" (UniqueName: \"kubernetes.io/projected/9197900b-6e4b-49c9-a41c-71fa58933754-kube-api-access-lgg6c\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.176978 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-client-ca\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.177014 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-config\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.177049 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9197900b-6e4b-49c9-a41c-71fa58933754-tmp\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.177083 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9197900b-6e4b-49c9-a41c-71fa58933754-serving-cert\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.177387 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-client-ca" (OuterVolumeSpecName: "client-ca") pod "ac38b6e9-af77-49d5-bb5b-4341f66d3224" (UID: "ac38b6e9-af77-49d5-bb5b-4341f66d3224"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.177685 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac38b6e9-af77-49d5-bb5b-4341f66d3224-tmp" (OuterVolumeSpecName: "tmp") pod "ac38b6e9-af77-49d5-bb5b-4341f66d3224" (UID: "ac38b6e9-af77-49d5-bb5b-4341f66d3224"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.178148 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-config" (OuterVolumeSpecName: "config") pod "ac38b6e9-af77-49d5-bb5b-4341f66d3224" (UID: "ac38b6e9-af77-49d5-bb5b-4341f66d3224"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.185057 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac38b6e9-af77-49d5-bb5b-4341f66d3224-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ac38b6e9-af77-49d5-bb5b-4341f66d3224" (UID: "ac38b6e9-af77-49d5-bb5b-4341f66d3224"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.191209 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac38b6e9-af77-49d5-bb5b-4341f66d3224-kube-api-access-vcvsc" (OuterVolumeSpecName: "kube-api-access-vcvsc") pod "ac38b6e9-af77-49d5-bb5b-4341f66d3224" (UID: "ac38b6e9-af77-49d5-bb5b-4341f66d3224"). InnerVolumeSpecName "kube-api-access-vcvsc". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.270964 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-12-crc"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.279017 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lgg6c\" (UniqueName: \"kubernetes.io/projected/9197900b-6e4b-49c9-a41c-71fa58933754-kube-api-access-lgg6c\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.279104 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-client-ca\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.279143 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-config\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.279179 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9197900b-6e4b-49c9-a41c-71fa58933754-tmp\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.279221 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9197900b-6e4b-49c9-a41c-71fa58933754-serving-cert\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.279231 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-12-crc"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.279427 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.280539 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9197900b-6e4b-49c9-a41c-71fa58933754-tmp\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.280553 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-client-ca\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.280615 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vcvsc\" (UniqueName: \"kubernetes.io/projected/ac38b6e9-af77-49d5-bb5b-4341f66d3224-kube-api-access-vcvsc\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.280652 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ac38b6e9-af77-49d5-bb5b-4341f66d3224-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.280665 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38b6e9-af77-49d5-bb5b-4341f66d3224-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.280678 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.280689 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac38b6e9-af77-49d5-bb5b-4341f66d3224-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.280886 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-config\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.283817 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver\"/\"installer-sa-dockercfg-bqqnb\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.284063 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver\"/\"kube-root-ca.crt\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.286830 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9197900b-6e4b-49c9-a41c-71fa58933754-serving-cert\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.301358 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgg6c\" (UniqueName: \"kubernetes.io/projected/9197900b-6e4b-49c9-a41c-71fa58933754-kube-api-access-lgg6c\") pod \"route-controller-manager-56d5578d57-rxhbj\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.338023 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.370038 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-849d8fbc49-rwrhg"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.370741 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c0c28656-8b42-4beb-be05-b5c0ee7cdca4" containerName="controller-manager" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.370762 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0c28656-8b42-4beb-be05-b5c0ee7cdca4" containerName="controller-manager" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.370870 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="c0c28656-8b42-4beb-be05-b5c0ee7cdca4" containerName="controller-manager" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.380704 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.381835 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-serving-cert\") pod \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.381863 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-config\") pod \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.381893 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-client-ca\") pod \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.382002 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-tmp\") pod \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.382077 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8z4b\" (UniqueName: \"kubernetes.io/projected/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-kube-api-access-f8z4b\") pod \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.382106 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-proxy-ca-bundles\") pod \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\" (UID: \"c0c28656-8b42-4beb-be05-b5c0ee7cdca4\") " Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.382267 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccef2781-57e5-49d7-ac44-70916671008c-kubelet-dir\") pod \"revision-pruner-12-crc\" (UID: \"ccef2781-57e5-49d7-ac44-70916671008c\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.382314 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccef2781-57e5-49d7-ac44-70916671008c-kube-api-access\") pod \"revision-pruner-12-crc\" (UID: \"ccef2781-57e5-49d7-ac44-70916671008c\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.383682 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-config" (OuterVolumeSpecName: "config") pod "c0c28656-8b42-4beb-be05-b5c0ee7cdca4" (UID: "c0c28656-8b42-4beb-be05-b5c0ee7cdca4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.384126 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c0c28656-8b42-4beb-be05-b5c0ee7cdca4" (UID: "c0c28656-8b42-4beb-be05-b5c0ee7cdca4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.384240 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-tmp" (OuterVolumeSpecName: "tmp") pod "c0c28656-8b42-4beb-be05-b5c0ee7cdca4" (UID: "c0c28656-8b42-4beb-be05-b5c0ee7cdca4"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.384391 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-client-ca" (OuterVolumeSpecName: "client-ca") pod "c0c28656-8b42-4beb-be05-b5c0ee7cdca4" (UID: "c0c28656-8b42-4beb-be05-b5c0ee7cdca4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.387469 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c0c28656-8b42-4beb-be05-b5c0ee7cdca4" (UID: "c0c28656-8b42-4beb-be05-b5c0ee7cdca4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.388078 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-kube-api-access-f8z4b" (OuterVolumeSpecName: "kube-api-access-f8z4b") pod "c0c28656-8b42-4beb-be05-b5c0ee7cdca4" (UID: "c0c28656-8b42-4beb-be05-b5c0ee7cdca4"). InnerVolumeSpecName "kube-api-access-f8z4b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.389141 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-849d8fbc49-rwrhg"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.420319 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483407 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-client-ca\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483464 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/3f455010-b074-48f0-8a8d-69a031c7cb03-tmp\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483509 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-config\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483548 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f455010-b074-48f0-8a8d-69a031c7cb03-serving-cert\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483592 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccef2781-57e5-49d7-ac44-70916671008c-kubelet-dir\") pod \"revision-pruner-12-crc\" (UID: \"ccef2781-57e5-49d7-ac44-70916671008c\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483625 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-proxy-ca-bundles\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483666 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccef2781-57e5-49d7-ac44-70916671008c-kube-api-access\") pod \"revision-pruner-12-crc\" (UID: \"ccef2781-57e5-49d7-ac44-70916671008c\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483709 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6phs\" (UniqueName: \"kubernetes.io/projected/3f455010-b074-48f0-8a8d-69a031c7cb03-kube-api-access-g6phs\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483764 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-f8z4b\" (UniqueName: \"kubernetes.io/projected/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-kube-api-access-f8z4b\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483780 5117 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483793 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483806 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483817 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483831 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0c28656-8b42-4beb-be05-b5c0ee7cdca4-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.483945 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccef2781-57e5-49d7-ac44-70916671008c-kubelet-dir\") pod \"revision-pruner-12-crc\" (UID: \"ccef2781-57e5-49d7-ac44-70916671008c\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.520047 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccef2781-57e5-49d7-ac44-70916671008c-kube-api-access\") pod \"revision-pruner-12-crc\" (UID: \"ccef2781-57e5-49d7-ac44-70916671008c\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.586199 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-client-ca\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.586273 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/3f455010-b074-48f0-8a8d-69a031c7cb03-tmp\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.586300 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-config\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.586327 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f455010-b074-48f0-8a8d-69a031c7cb03-serving-cert\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.586358 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-proxy-ca-bundles\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.586396 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g6phs\" (UniqueName: \"kubernetes.io/projected/3f455010-b074-48f0-8a8d-69a031c7cb03-kube-api-access-g6phs\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.587872 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-client-ca\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.588584 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/3f455010-b074-48f0-8a8d-69a031c7cb03-tmp\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.588870 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-proxy-ca-bundles\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.591056 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-config\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.592905 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f455010-b074-48f0-8a8d-69a031c7cb03-serving-cert\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.606229 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6phs\" (UniqueName: \"kubernetes.io/projected/3f455010-b074-48f0-8a8d-69a031c7cb03-kube-api-access-g6phs\") pod \"controller-manager-849d8fbc49-rwrhg\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.647689 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.710533 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.898574 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.905458 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-849d8fbc49-rwrhg"] Mar 21 00:13:59 crc kubenswrapper[5117]: W0321 00:13:59.912968 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f455010_b074_48f0_8a8d_69a031c7cb03.slice/crio-245bf6628706d0f157389b66aa6abc0aea782cdc007e4f8729015677d422bea1 WatchSource:0}: Error finding container 245bf6628706d0f157389b66aa6abc0aea782cdc007e4f8729015677d422bea1: Status 404 returned error can't find the container with id 245bf6628706d0f157389b66aa6abc0aea782cdc007e4f8729015677d422bea1 Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.929435 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" event={"ID":"c0c28656-8b42-4beb-be05-b5c0ee7cdca4","Type":"ContainerDied","Data":"7648e6dbc8de7fada85d45f9f1cedd8bd90573bab102ee64ec07c10789567066"} Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.929493 5117 scope.go:117] "RemoveContainer" containerID="f58bd6065bdfd1b3bf07376d7d094f3ad1ead1ad9ba24fc20d100db0e091587d" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.929661 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b449fd8cc-9ckff" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.934998 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" event={"ID":"3f455010-b074-48f0-8a8d-69a031c7cb03","Type":"ContainerStarted","Data":"245bf6628706d0f157389b66aa6abc0aea782cdc007e4f8729015677d422bea1"} Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.936318 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" event={"ID":"9197900b-6e4b-49c9-a41c-71fa58933754","Type":"ContainerStarted","Data":"07fe3590d8d1b555ca5b81cb7bf502b8cc0af498577f9cbeed0af874f59c0f88"} Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.937926 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" event={"ID":"ac38b6e9-af77-49d5-bb5b-4341f66d3224","Type":"ContainerDied","Data":"09f3f77e2be372aa234c97865bc71ceca91acb856e31f9dfd514f83b862de9bd"} Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.938006 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db77b7694-njxld" Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.946593 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b449fd8cc-9ckff"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.949120 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7b449fd8cc-9ckff"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.956756 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db77b7694-njxld"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.959128 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db77b7694-njxld"] Mar 21 00:13:59 crc kubenswrapper[5117]: I0321 00:13:59.959271 5117 scope.go:117] "RemoveContainer" containerID="874da865045fae9905bc80bfe3ae82b9e275936a1f84be89734dfa4c6794e6ae" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.047513 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-12-crc"] Mar 21 00:14:00 crc kubenswrapper[5117]: W0321 00:14:00.055534 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podccef2781_57e5_49d7_ac44_70916671008c.slice/crio-c1123665f6486c7867680b484a5713dc7391c931323b4f2fd6773fc460ae28be WatchSource:0}: Error finding container c1123665f6486c7867680b484a5713dc7391c931323b4f2fd6773fc460ae28be: Status 404 returned error can't find the container with id c1123665f6486c7867680b484a5713dc7391c931323b4f2fd6773fc460ae28be Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.125631 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567534-vhz59"] Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.129574 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567534-vhz59" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.133389 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567534-vhz59"] Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.135662 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.136104 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.136213 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.195794 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz2d2\" (UniqueName: \"kubernetes.io/projected/ba892113-95e3-461f-a9c3-34295e210034-kube-api-access-sz2d2\") pod \"auto-csr-approver-29567534-vhz59\" (UID: \"ba892113-95e3-461f-a9c3-34295e210034\") " pod="openshift-infra/auto-csr-approver-29567534-vhz59" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.297244 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sz2d2\" (UniqueName: \"kubernetes.io/projected/ba892113-95e3-461f-a9c3-34295e210034-kube-api-access-sz2d2\") pod \"auto-csr-approver-29567534-vhz59\" (UID: \"ba892113-95e3-461f-a9c3-34295e210034\") " pod="openshift-infra/auto-csr-approver-29567534-vhz59" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.329639 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz2d2\" (UniqueName: \"kubernetes.io/projected/ba892113-95e3-461f-a9c3-34295e210034-kube-api-access-sz2d2\") pod \"auto-csr-approver-29567534-vhz59\" (UID: \"ba892113-95e3-461f-a9c3-34295e210034\") " pod="openshift-infra/auto-csr-approver-29567534-vhz59" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.353260 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mz5qp"] Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.353559 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mz5qp" podUID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerName="registry-server" containerID="cri-o://0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe" gracePeriod=2 Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.456113 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567534-vhz59" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.738939 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.740409 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.787015 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.806980 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrlnz\" (UniqueName: \"kubernetes.io/projected/30a52c1e-3bfa-4789-a6ab-d851d6139934-kube-api-access-lrlnz\") pod \"30a52c1e-3bfa-4789-a6ab-d851d6139934\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.807210 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-utilities\") pod \"30a52c1e-3bfa-4789-a6ab-d851d6139934\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.807262 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-catalog-content\") pod \"30a52c1e-3bfa-4789-a6ab-d851d6139934\" (UID: \"30a52c1e-3bfa-4789-a6ab-d851d6139934\") " Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.809591 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-utilities" (OuterVolumeSpecName: "utilities") pod "30a52c1e-3bfa-4789-a6ab-d851d6139934" (UID: "30a52c1e-3bfa-4789-a6ab-d851d6139934"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.817358 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a52c1e-3bfa-4789-a6ab-d851d6139934-kube-api-access-lrlnz" (OuterVolumeSpecName: "kube-api-access-lrlnz") pod "30a52c1e-3bfa-4789-a6ab-d851d6139934" (UID: "30a52c1e-3bfa-4789-a6ab-d851d6139934"). InnerVolumeSpecName "kube-api-access-lrlnz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.872230 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30a52c1e-3bfa-4789-a6ab-d851d6139934" (UID: "30a52c1e-3bfa-4789-a6ab-d851d6139934"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.902134 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567534-vhz59"] Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.909147 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.909179 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a52c1e-3bfa-4789-a6ab-d851d6139934-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.909189 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lrlnz\" (UniqueName: \"kubernetes.io/projected/30a52c1e-3bfa-4789-a6ab-d851d6139934-kube-api-access-lrlnz\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:00 crc kubenswrapper[5117]: W0321 00:14:00.909742 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba892113_95e3_461f_a9c3_34295e210034.slice/crio-6d824532f52457ffbbfebd9c21aedb4b1b041caf03412957b3d4bd041fb4195a WatchSource:0}: Error finding container 6d824532f52457ffbbfebd9c21aedb4b1b041caf03412957b3d4bd041fb4195a: Status 404 returned error can't find the container with id 6d824532f52457ffbbfebd9c21aedb4b1b041caf03412957b3d4bd041fb4195a Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.945702 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"ccef2781-57e5-49d7-ac44-70916671008c","Type":"ContainerStarted","Data":"722d490055aa7dce9bd57804af9b4eb8d1a1b465c0eaad425e4b0dfa261d2ff7"} Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.945751 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"ccef2781-57e5-49d7-ac44-70916671008c","Type":"ContainerStarted","Data":"c1123665f6486c7867680b484a5713dc7391c931323b4f2fd6773fc460ae28be"} Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.961216 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" event={"ID":"3f455010-b074-48f0-8a8d-69a031c7cb03","Type":"ContainerStarted","Data":"241eca1a9d4116785e14f08fca13392092ca7ca35024957b14d0f196e9e922ac"} Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.961713 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.965772 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" event={"ID":"9197900b-6e4b-49c9-a41c-71fa58933754","Type":"ContainerStarted","Data":"44f26c8620777f66cdb2d06a6921197264f8e040030ec4f6478c45c5977c6dc7"} Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.967081 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.974096 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.974817 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-12-crc" podStartSLOduration=1.974791675 podStartE2EDuration="1.974791675s" podCreationTimestamp="2026-03-21 00:13:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:14:00.974192947 +0000 UTC m=+154.267480129" watchObservedRunningTime="2026-03-21 00:14:00.974791675 +0000 UTC m=+154.268078847" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.975420 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.977350 5117 generic.go:358] "Generic (PLEG): container finished" podID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerID="0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe" exitCode=0 Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.977427 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mz5qp" event={"ID":"30a52c1e-3bfa-4789-a6ab-d851d6139934","Type":"ContainerDied","Data":"0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe"} Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.977468 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mz5qp" event={"ID":"30a52c1e-3bfa-4789-a6ab-d851d6139934","Type":"ContainerDied","Data":"bda23d6d7e0f51eba0cc6fc3de72270405282e5fae544c3a9ad2d92125979633"} Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.977497 5117 scope.go:117] "RemoveContainer" containerID="0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.977692 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mz5qp" Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.979437 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567534-vhz59" event={"ID":"ba892113-95e3-461f-a9c3-34295e210034","Type":"ContainerStarted","Data":"6d824532f52457ffbbfebd9c21aedb4b1b041caf03412957b3d4bd041fb4195a"} Mar 21 00:14:00 crc kubenswrapper[5117]: I0321 00:14:00.994145 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" podStartSLOduration=2.994122891 podStartE2EDuration="2.994122891s" podCreationTimestamp="2026-03-21 00:13:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:14:00.99344469 +0000 UTC m=+154.286731882" watchObservedRunningTime="2026-03-21 00:14:00.994122891 +0000 UTC m=+154.287410063" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.002498 5117 scope.go:117] "RemoveContainer" containerID="3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.016385 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" podStartSLOduration=3.016368237 podStartE2EDuration="3.016368237s" podCreationTimestamp="2026-03-21 00:13:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:14:01.0151781 +0000 UTC m=+154.308465272" watchObservedRunningTime="2026-03-21 00:14:01.016368237 +0000 UTC m=+154.309655409" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.068217 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mz5qp"] Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.068300 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mz5qp"] Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.073986 5117 scope.go:117] "RemoveContainer" containerID="ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.133563 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.141989 5117 scope.go:117] "RemoveContainer" containerID="0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe" Mar 21 00:14:01 crc kubenswrapper[5117]: E0321 00:14:01.142597 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe\": container with ID starting with 0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe not found: ID does not exist" containerID="0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.142664 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe"} err="failed to get container status \"0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe\": rpc error: code = NotFound desc = could not find container \"0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe\": container with ID starting with 0e6cab1dfea37ab8bd17400fb86190a7989d94782fd7200064f9c6a58e2250fe not found: ID does not exist" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.142727 5117 scope.go:117] "RemoveContainer" containerID="3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8" Mar 21 00:14:01 crc kubenswrapper[5117]: E0321 00:14:01.143156 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8\": container with ID starting with 3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8 not found: ID does not exist" containerID="3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.143202 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8"} err="failed to get container status \"3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8\": rpc error: code = NotFound desc = could not find container \"3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8\": container with ID starting with 3d94686129d7a6f59ea6c1145363b2f5cfc816d94119589d0a3aa7cd344b7ef8 not found: ID does not exist" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.143220 5117 scope.go:117] "RemoveContainer" containerID="ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae" Mar 21 00:14:01 crc kubenswrapper[5117]: E0321 00:14:01.143858 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae\": container with ID starting with ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae not found: ID does not exist" containerID="ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.143885 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae"} err="failed to get container status \"ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae\": rpc error: code = NotFound desc = could not find container \"ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae\": container with ID starting with ddf1917921bf59b81a3fc7a3b8a2db2f601cd459d329dcfd6276db3784bc98ae not found: ID does not exist" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.172013 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.586039 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a52c1e-3bfa-4789-a6ab-d851d6139934" path="/var/lib/kubelet/pods/30a52c1e-3bfa-4789-a6ab-d851d6139934/volumes" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.586965 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac38b6e9-af77-49d5-bb5b-4341f66d3224" path="/var/lib/kubelet/pods/ac38b6e9-af77-49d5-bb5b-4341f66d3224/volumes" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.587732 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0c28656-8b42-4beb-be05-b5c0ee7cdca4" path="/var/lib/kubelet/pods/c0c28656-8b42-4beb-be05-b5c0ee7cdca4/volumes" Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.986842 5117 generic.go:358] "Generic (PLEG): container finished" podID="ccef2781-57e5-49d7-ac44-70916671008c" containerID="722d490055aa7dce9bd57804af9b4eb8d1a1b465c0eaad425e4b0dfa261d2ff7" exitCode=0 Mar 21 00:14:01 crc kubenswrapper[5117]: I0321 00:14:01.986972 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"ccef2781-57e5-49d7-ac44-70916671008c","Type":"ContainerDied","Data":"722d490055aa7dce9bd57804af9b4eb8d1a1b465c0eaad425e4b0dfa261d2ff7"} Mar 21 00:14:03 crc kubenswrapper[5117]: I0321 00:14:03.281455 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:14:03 crc kubenswrapper[5117]: I0321 00:14:03.410654 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccef2781-57e5-49d7-ac44-70916671008c-kube-api-access\") pod \"ccef2781-57e5-49d7-ac44-70916671008c\" (UID: \"ccef2781-57e5-49d7-ac44-70916671008c\") " Mar 21 00:14:03 crc kubenswrapper[5117]: I0321 00:14:03.410754 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccef2781-57e5-49d7-ac44-70916671008c-kubelet-dir\") pod \"ccef2781-57e5-49d7-ac44-70916671008c\" (UID: \"ccef2781-57e5-49d7-ac44-70916671008c\") " Mar 21 00:14:03 crc kubenswrapper[5117]: I0321 00:14:03.411064 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccef2781-57e5-49d7-ac44-70916671008c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ccef2781-57e5-49d7-ac44-70916671008c" (UID: "ccef2781-57e5-49d7-ac44-70916671008c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:14:03 crc kubenswrapper[5117]: I0321 00:14:03.424804 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccef2781-57e5-49d7-ac44-70916671008c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ccef2781-57e5-49d7-ac44-70916671008c" (UID: "ccef2781-57e5-49d7-ac44-70916671008c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:03 crc kubenswrapper[5117]: I0321 00:14:03.521491 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ccef2781-57e5-49d7-ac44-70916671008c-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:03 crc kubenswrapper[5117]: I0321 00:14:03.521530 5117 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ccef2781-57e5-49d7-ac44-70916671008c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:04 crc kubenswrapper[5117]: I0321 00:14:04.013853 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"ccef2781-57e5-49d7-ac44-70916671008c","Type":"ContainerDied","Data":"c1123665f6486c7867680b484a5713dc7391c931323b4f2fd6773fc460ae28be"} Mar 21 00:14:04 crc kubenswrapper[5117]: I0321 00:14:04.014153 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1123665f6486c7867680b484a5713dc7391c931323b4f2fd6773fc460ae28be" Mar 21 00:14:04 crc kubenswrapper[5117]: I0321 00:14:04.013860 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Mar 21 00:14:04 crc kubenswrapper[5117]: I0321 00:14:04.379454 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:14:04 crc kubenswrapper[5117]: I0321 00:14:04.753387 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qz57f"] Mar 21 00:14:04 crc kubenswrapper[5117]: I0321 00:14:04.753772 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qz57f" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerName="registry-server" containerID="cri-o://8bfa5e41be23006122c4f02a480d001bccc38bb00d52d38a26ee119322b434b9" gracePeriod=2 Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.027158 5117 generic.go:358] "Generic (PLEG): container finished" podID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerID="8bfa5e41be23006122c4f02a480d001bccc38bb00d52d38a26ee119322b434b9" exitCode=0 Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.027375 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz57f" event={"ID":"c15a6f5e-ffea-444b-98d1-3922aeb22128","Type":"ContainerDied","Data":"8bfa5e41be23006122c4f02a480d001bccc38bb00d52d38a26ee119322b434b9"} Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.751568 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.855961 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvw2v\" (UniqueName: \"kubernetes.io/projected/c15a6f5e-ffea-444b-98d1-3922aeb22128-kube-api-access-kvw2v\") pod \"c15a6f5e-ffea-444b-98d1-3922aeb22128\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.856179 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-catalog-content\") pod \"c15a6f5e-ffea-444b-98d1-3922aeb22128\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.856436 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-utilities\") pod \"c15a6f5e-ffea-444b-98d1-3922aeb22128\" (UID: \"c15a6f5e-ffea-444b-98d1-3922aeb22128\") " Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.857494 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-utilities" (OuterVolumeSpecName: "utilities") pod "c15a6f5e-ffea-444b-98d1-3922aeb22128" (UID: "c15a6f5e-ffea-444b-98d1-3922aeb22128"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.871474 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15a6f5e-ffea-444b-98d1-3922aeb22128-kube-api-access-kvw2v" (OuterVolumeSpecName: "kube-api-access-kvw2v") pod "c15a6f5e-ffea-444b-98d1-3922aeb22128" (UID: "c15a6f5e-ffea-444b-98d1-3922aeb22128"). InnerVolumeSpecName "kube-api-access-kvw2v". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.958488 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.958530 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kvw2v\" (UniqueName: \"kubernetes.io/projected/c15a6f5e-ffea-444b-98d1-3922aeb22128-kube-api-access-kvw2v\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:05 crc kubenswrapper[5117]: I0321 00:14:05.977235 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c15a6f5e-ffea-444b-98d1-3922aeb22128" (UID: "c15a6f5e-ffea-444b-98d1-3922aeb22128"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.035909 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz57f" Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.036004 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz57f" event={"ID":"c15a6f5e-ffea-444b-98d1-3922aeb22128","Type":"ContainerDied","Data":"b4de5e9e85dbe09789e7ae5965256793925d7d960eb4993bc87034c56df7273a"} Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.036057 5117 scope.go:117] "RemoveContainer" containerID="8bfa5e41be23006122c4f02a480d001bccc38bb00d52d38a26ee119322b434b9" Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.038136 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567534-vhz59" event={"ID":"ba892113-95e3-461f-a9c3-34295e210034","Type":"ContainerStarted","Data":"5789968292dbaba1a6ec24956d0efe67faa98c185c816821c00fef9e8009bec5"} Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.050512 5117 scope.go:117] "RemoveContainer" containerID="61db64811d8fa48f4dea7bc5e6b24139fa0f2fa85efb6e7f6637840a6d8953e9" Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.060243 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15a6f5e-ffea-444b-98d1-3922aeb22128-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.066764 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qz57f"] Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.069753 5117 scope.go:117] "RemoveContainer" containerID="d1ec1e95e3c26d54fc4427b668575182a6839560c78afd57090e9f962248df2f" Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.076338 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qz57f"] Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.578736 5117 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-2xhbb" Mar 21 00:14:06 crc kubenswrapper[5117]: I0321 00:14:06.591450 5117 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-2xhbb" Mar 21 00:14:07 crc kubenswrapper[5117]: I0321 00:14:07.045522 5117 generic.go:358] "Generic (PLEG): container finished" podID="ba892113-95e3-461f-a9c3-34295e210034" containerID="5789968292dbaba1a6ec24956d0efe67faa98c185c816821c00fef9e8009bec5" exitCode=0 Mar 21 00:14:07 crc kubenswrapper[5117]: I0321 00:14:07.045600 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567534-vhz59" event={"ID":"ba892113-95e3-461f-a9c3-34295e210034","Type":"ContainerDied","Data":"5789968292dbaba1a6ec24956d0efe67faa98c185c816821c00fef9e8009bec5"} Mar 21 00:14:07 crc kubenswrapper[5117]: I0321 00:14:07.592339 5117 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2026-04-20 00:09:06 +0000 UTC" deadline="2026-04-11 18:36:32.987158797 +0000 UTC" Mar 21 00:14:07 crc kubenswrapper[5117]: I0321 00:14:07.592404 5117 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="522h22m25.394758356s" Mar 21 00:14:07 crc kubenswrapper[5117]: I0321 00:14:07.599054 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" path="/var/lib/kubelet/pods/c15a6f5e-ffea-444b-98d1-3922aeb22128/volumes" Mar 21 00:14:08 crc kubenswrapper[5117]: I0321 00:14:08.339698 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567534-vhz59" Mar 21 00:14:08 crc kubenswrapper[5117]: I0321 00:14:08.400856 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz2d2\" (UniqueName: \"kubernetes.io/projected/ba892113-95e3-461f-a9c3-34295e210034-kube-api-access-sz2d2\") pod \"ba892113-95e3-461f-a9c3-34295e210034\" (UID: \"ba892113-95e3-461f-a9c3-34295e210034\") " Mar 21 00:14:08 crc kubenswrapper[5117]: I0321 00:14:08.406554 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba892113-95e3-461f-a9c3-34295e210034-kube-api-access-sz2d2" (OuterVolumeSpecName: "kube-api-access-sz2d2") pod "ba892113-95e3-461f-a9c3-34295e210034" (UID: "ba892113-95e3-461f-a9c3-34295e210034"). InnerVolumeSpecName "kube-api-access-sz2d2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:08 crc kubenswrapper[5117]: I0321 00:14:08.502074 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-sz2d2\" (UniqueName: \"kubernetes.io/projected/ba892113-95e3-461f-a9c3-34295e210034-kube-api-access-sz2d2\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:08 crc kubenswrapper[5117]: I0321 00:14:08.593314 5117 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2026-04-20 00:09:06 +0000 UTC" deadline="2026-04-11 12:55:59.064398224 +0000 UTC" Mar 21 00:14:08 crc kubenswrapper[5117]: I0321 00:14:08.593361 5117 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="516h41m50.471041876s" Mar 21 00:14:09 crc kubenswrapper[5117]: I0321 00:14:09.059995 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567534-vhz59" event={"ID":"ba892113-95e3-461f-a9c3-34295e210034","Type":"ContainerDied","Data":"6d824532f52457ffbbfebd9c21aedb4b1b041caf03412957b3d4bd041fb4195a"} Mar 21 00:14:09 crc kubenswrapper[5117]: I0321 00:14:09.060031 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d824532f52457ffbbfebd9c21aedb4b1b041caf03412957b3d4bd041fb4195a" Mar 21 00:14:09 crc kubenswrapper[5117]: I0321 00:14:09.060043 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567534-vhz59" Mar 21 00:14:17 crc kubenswrapper[5117]: I0321 00:14:17.966017 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-849d8fbc49-rwrhg"] Mar 21 00:14:17 crc kubenswrapper[5117]: I0321 00:14:17.966898 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" podUID="3f455010-b074-48f0-8a8d-69a031c7cb03" containerName="controller-manager" containerID="cri-o://241eca1a9d4116785e14f08fca13392092ca7ca35024957b14d0f196e9e922ac" gracePeriod=30 Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.001447 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj"] Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.001809 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" podUID="9197900b-6e4b-49c9-a41c-71fa58933754" containerName="route-controller-manager" containerID="cri-o://44f26c8620777f66cdb2d06a6921197264f8e040030ec4f6478c45c5977c6dc7" gracePeriod=30 Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.141229 5117 generic.go:358] "Generic (PLEG): container finished" podID="3f455010-b074-48f0-8a8d-69a031c7cb03" containerID="241eca1a9d4116785e14f08fca13392092ca7ca35024957b14d0f196e9e922ac" exitCode=0 Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.141766 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" event={"ID":"3f455010-b074-48f0-8a8d-69a031c7cb03","Type":"ContainerDied","Data":"241eca1a9d4116785e14f08fca13392092ca7ca35024957b14d0f196e9e922ac"} Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.148114 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" event={"ID":"9197900b-6e4b-49c9-a41c-71fa58933754","Type":"ContainerDied","Data":"44f26c8620777f66cdb2d06a6921197264f8e040030ec4f6478c45c5977c6dc7"} Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.148368 5117 generic.go:358] "Generic (PLEG): container finished" podID="9197900b-6e4b-49c9-a41c-71fa58933754" containerID="44f26c8620777f66cdb2d06a6921197264f8e040030ec4f6478c45c5977c6dc7" exitCode=0 Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.540128 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571215 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj"] Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571769 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerName="extract-utilities" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571787 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerName="extract-utilities" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571800 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerName="registry-server" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571807 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerName="registry-server" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571814 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ba892113-95e3-461f-a9c3-34295e210034" containerName="oc" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571820 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba892113-95e3-461f-a9c3-34295e210034" containerName="oc" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571829 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerName="registry-server" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571834 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerName="registry-server" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571854 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9197900b-6e4b-49c9-a41c-71fa58933754" containerName="route-controller-manager" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571860 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="9197900b-6e4b-49c9-a41c-71fa58933754" containerName="route-controller-manager" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571869 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerName="extract-content" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571874 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerName="extract-content" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571884 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ccef2781-57e5-49d7-ac44-70916671008c" containerName="pruner" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571889 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccef2781-57e5-49d7-ac44-70916671008c" containerName="pruner" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571901 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerName="extract-utilities" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571907 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerName="extract-utilities" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571920 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerName="extract-content" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.571927 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerName="extract-content" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.572012 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="9197900b-6e4b-49c9-a41c-71fa58933754" containerName="route-controller-manager" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.572022 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ba892113-95e3-461f-a9c3-34295e210034" containerName="oc" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.572031 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="c15a6f5e-ffea-444b-98d1-3922aeb22128" containerName="registry-server" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.572039 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ccef2781-57e5-49d7-ac44-70916671008c" containerName="pruner" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.572048 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="30a52c1e-3bfa-4789-a6ab-d851d6139934" containerName="registry-server" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.578853 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.587967 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj"] Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.636037 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9197900b-6e4b-49c9-a41c-71fa58933754-tmp\") pod \"9197900b-6e4b-49c9-a41c-71fa58933754\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.636172 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-client-ca\") pod \"9197900b-6e4b-49c9-a41c-71fa58933754\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.636309 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgg6c\" (UniqueName: \"kubernetes.io/projected/9197900b-6e4b-49c9-a41c-71fa58933754-kube-api-access-lgg6c\") pod \"9197900b-6e4b-49c9-a41c-71fa58933754\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.636336 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-config\") pod \"9197900b-6e4b-49c9-a41c-71fa58933754\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.636602 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9197900b-6e4b-49c9-a41c-71fa58933754-tmp" (OuterVolumeSpecName: "tmp") pod "9197900b-6e4b-49c9-a41c-71fa58933754" (UID: "9197900b-6e4b-49c9-a41c-71fa58933754"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.636903 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-client-ca" (OuterVolumeSpecName: "client-ca") pod "9197900b-6e4b-49c9-a41c-71fa58933754" (UID: "9197900b-6e4b-49c9-a41c-71fa58933754"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.636978 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-config" (OuterVolumeSpecName: "config") pod "9197900b-6e4b-49c9-a41c-71fa58933754" (UID: "9197900b-6e4b-49c9-a41c-71fa58933754"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.637027 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9197900b-6e4b-49c9-a41c-71fa58933754-serving-cert\") pod \"9197900b-6e4b-49c9-a41c-71fa58933754\" (UID: \"9197900b-6e4b-49c9-a41c-71fa58933754\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.637623 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-client-ca\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.637689 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76e297ff-c558-4d61-8d10-6ce8a183f450-serving-cert\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.637720 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/76e297ff-c558-4d61-8d10-6ce8a183f450-tmp\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.637768 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-config\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.637790 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxr67\" (UniqueName: \"kubernetes.io/projected/76e297ff-c558-4d61-8d10-6ce8a183f450-kube-api-access-pxr67\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.637860 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.637872 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9197900b-6e4b-49c9-a41c-71fa58933754-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.637880 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9197900b-6e4b-49c9-a41c-71fa58933754-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.643015 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9197900b-6e4b-49c9-a41c-71fa58933754-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9197900b-6e4b-49c9-a41c-71fa58933754" (UID: "9197900b-6e4b-49c9-a41c-71fa58933754"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.643692 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9197900b-6e4b-49c9-a41c-71fa58933754-kube-api-access-lgg6c" (OuterVolumeSpecName: "kube-api-access-lgg6c") pod "9197900b-6e4b-49c9-a41c-71fa58933754" (UID: "9197900b-6e4b-49c9-a41c-71fa58933754"). InnerVolumeSpecName "kube-api-access-lgg6c". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.678891 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.703371 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b468689cd-889sb"] Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.704197 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3f455010-b074-48f0-8a8d-69a031c7cb03" containerName="controller-manager" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.704226 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f455010-b074-48f0-8a8d-69a031c7cb03" containerName="controller-manager" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.704378 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3f455010-b074-48f0-8a8d-69a031c7cb03" containerName="controller-manager" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.709806 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.713217 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b468689cd-889sb"] Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739164 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-client-ca\") pod \"3f455010-b074-48f0-8a8d-69a031c7cb03\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739215 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/3f455010-b074-48f0-8a8d-69a031c7cb03-tmp\") pod \"3f455010-b074-48f0-8a8d-69a031c7cb03\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739240 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f455010-b074-48f0-8a8d-69a031c7cb03-serving-cert\") pod \"3f455010-b074-48f0-8a8d-69a031c7cb03\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739365 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6phs\" (UniqueName: \"kubernetes.io/projected/3f455010-b074-48f0-8a8d-69a031c7cb03-kube-api-access-g6phs\") pod \"3f455010-b074-48f0-8a8d-69a031c7cb03\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739455 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-config\") pod \"3f455010-b074-48f0-8a8d-69a031c7cb03\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739542 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-proxy-ca-bundles\") pod \"3f455010-b074-48f0-8a8d-69a031c7cb03\" (UID: \"3f455010-b074-48f0-8a8d-69a031c7cb03\") " Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739715 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76e297ff-c558-4d61-8d10-6ce8a183f450-serving-cert\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739753 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/76e297ff-c558-4d61-8d10-6ce8a183f450-tmp\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739784 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-config\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739801 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pxr67\" (UniqueName: \"kubernetes.io/projected/76e297ff-c558-4d61-8d10-6ce8a183f450-kube-api-access-pxr67\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739852 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-client-ca\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739853 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f455010-b074-48f0-8a8d-69a031c7cb03-tmp" (OuterVolumeSpecName: "tmp") pod "3f455010-b074-48f0-8a8d-69a031c7cb03" (UID: "3f455010-b074-48f0-8a8d-69a031c7cb03"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739937 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9197900b-6e4b-49c9-a41c-71fa58933754-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.739955 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lgg6c\" (UniqueName: \"kubernetes.io/projected/9197900b-6e4b-49c9-a41c-71fa58933754-kube-api-access-lgg6c\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.740560 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/76e297ff-c558-4d61-8d10-6ce8a183f450-tmp\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.740619 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-config" (OuterVolumeSpecName: "config") pod "3f455010-b074-48f0-8a8d-69a031c7cb03" (UID: "3f455010-b074-48f0-8a8d-69a031c7cb03"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.741273 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-client-ca\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.741333 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-config\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.741728 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-client-ca" (OuterVolumeSpecName: "client-ca") pod "3f455010-b074-48f0-8a8d-69a031c7cb03" (UID: "3f455010-b074-48f0-8a8d-69a031c7cb03"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.742602 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3f455010-b074-48f0-8a8d-69a031c7cb03" (UID: "3f455010-b074-48f0-8a8d-69a031c7cb03"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.744520 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f455010-b074-48f0-8a8d-69a031c7cb03-kube-api-access-g6phs" (OuterVolumeSpecName: "kube-api-access-g6phs") pod "3f455010-b074-48f0-8a8d-69a031c7cb03" (UID: "3f455010-b074-48f0-8a8d-69a031c7cb03"). InnerVolumeSpecName "kube-api-access-g6phs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.744549 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f455010-b074-48f0-8a8d-69a031c7cb03-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3f455010-b074-48f0-8a8d-69a031c7cb03" (UID: "3f455010-b074-48f0-8a8d-69a031c7cb03"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.746304 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76e297ff-c558-4d61-8d10-6ce8a183f450-serving-cert\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.760391 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxr67\" (UniqueName: \"kubernetes.io/projected/76e297ff-c558-4d61-8d10-6ce8a183f450-kube-api-access-pxr67\") pod \"route-controller-manager-54646bb8cc-259gj\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.840854 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/764d6012-5a93-4388-8efd-4a38a439ead8-tmp\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.841698 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-config\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.841791 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-client-ca\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.841891 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-proxy-ca-bundles\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.842139 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/764d6012-5a93-4388-8efd-4a38a439ead8-serving-cert\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.842399 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxcwk\" (UniqueName: \"kubernetes.io/projected/764d6012-5a93-4388-8efd-4a38a439ead8-kube-api-access-kxcwk\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.842610 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.842635 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/3f455010-b074-48f0-8a8d-69a031c7cb03-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.842649 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f455010-b074-48f0-8a8d-69a031c7cb03-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.842663 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-g6phs\" (UniqueName: \"kubernetes.io/projected/3f455010-b074-48f0-8a8d-69a031c7cb03-kube-api-access-g6phs\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.842680 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.842694 5117 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f455010-b074-48f0-8a8d-69a031c7cb03-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.901100 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.943635 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-proxy-ca-bundles\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.943706 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/764d6012-5a93-4388-8efd-4a38a439ead8-serving-cert\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.943753 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kxcwk\" (UniqueName: \"kubernetes.io/projected/764d6012-5a93-4388-8efd-4a38a439ead8-kube-api-access-kxcwk\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.943788 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/764d6012-5a93-4388-8efd-4a38a439ead8-tmp\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.943843 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-config\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.943868 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-client-ca\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.944507 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/764d6012-5a93-4388-8efd-4a38a439ead8-tmp\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.944670 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-proxy-ca-bundles\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.945452 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-config\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.947176 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-client-ca\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.948952 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/764d6012-5a93-4388-8efd-4a38a439ead8-serving-cert\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:18 crc kubenswrapper[5117]: I0321 00:14:18.960507 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxcwk\" (UniqueName: \"kubernetes.io/projected/764d6012-5a93-4388-8efd-4a38a439ead8-kube-api-access-kxcwk\") pod \"controller-manager-b468689cd-889sb\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.026200 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.158133 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.158180 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-849d8fbc49-rwrhg" event={"ID":"3f455010-b074-48f0-8a8d-69a031c7cb03","Type":"ContainerDied","Data":"245bf6628706d0f157389b66aa6abc0aea782cdc007e4f8729015677d422bea1"} Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.158990 5117 scope.go:117] "RemoveContainer" containerID="241eca1a9d4116785e14f08fca13392092ca7ca35024957b14d0f196e9e922ac" Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.161382 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.161437 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj" event={"ID":"9197900b-6e4b-49c9-a41c-71fa58933754","Type":"ContainerDied","Data":"07fe3590d8d1b555ca5b81cb7bf502b8cc0af498577f9cbeed0af874f59c0f88"} Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.195978 5117 scope.go:117] "RemoveContainer" containerID="44f26c8620777f66cdb2d06a6921197264f8e040030ec4f6478c45c5977c6dc7" Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.201686 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-849d8fbc49-rwrhg"] Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.210130 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-849d8fbc49-rwrhg"] Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.212796 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj"] Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.215503 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d5578d57-rxhbj"] Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.323884 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj"] Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.420973 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b468689cd-889sb"] Mar 21 00:14:19 crc kubenswrapper[5117]: W0321 00:14:19.436881 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod764d6012_5a93_4388_8efd_4a38a439ead8.slice/crio-3aed14f0b93b294998031674e374e96b0fe60212879d072412f9b8b9ba4cc7ea WatchSource:0}: Error finding container 3aed14f0b93b294998031674e374e96b0fe60212879d072412f9b8b9ba4cc7ea: Status 404 returned error can't find the container with id 3aed14f0b93b294998031674e374e96b0fe60212879d072412f9b8b9ba4cc7ea Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.585721 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f455010-b074-48f0-8a8d-69a031c7cb03" path="/var/lib/kubelet/pods/3f455010-b074-48f0-8a8d-69a031c7cb03/volumes" Mar 21 00:14:19 crc kubenswrapper[5117]: I0321 00:14:19.586291 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9197900b-6e4b-49c9-a41c-71fa58933754" path="/var/lib/kubelet/pods/9197900b-6e4b-49c9-a41c-71fa58933754/volumes" Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.175484 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" event={"ID":"764d6012-5a93-4388-8efd-4a38a439ead8","Type":"ContainerStarted","Data":"337223f3388f64fbb81d010acbf676df8616a2b4bd222dfcafa40ce38936c177"} Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.175554 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" event={"ID":"764d6012-5a93-4388-8efd-4a38a439ead8","Type":"ContainerStarted","Data":"3aed14f0b93b294998031674e374e96b0fe60212879d072412f9b8b9ba4cc7ea"} Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.179490 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.181267 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" event={"ID":"76e297ff-c558-4d61-8d10-6ce8a183f450","Type":"ContainerStarted","Data":"458e8653f23e665e6dd547636b7a0658a50275e8833ce2c79735b1f16f30b72d"} Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.181294 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" event={"ID":"76e297ff-c558-4d61-8d10-6ce8a183f450","Type":"ContainerStarted","Data":"5ac5527c2a4aebf2addfba42598e9c48d1b3c101c5abea9f715bf73007dd1820"} Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.182158 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.187605 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.196936 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" podStartSLOduration=3.196911248 podStartE2EDuration="3.196911248s" podCreationTimestamp="2026-03-21 00:14:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:14:20.196609289 +0000 UTC m=+173.489896481" watchObservedRunningTime="2026-03-21 00:14:20.196911248 +0000 UTC m=+173.490198420" Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.213128 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" podStartSLOduration=2.213109218 podStartE2EDuration="2.213109218s" podCreationTimestamp="2026-03-21 00:14:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:14:20.211973853 +0000 UTC m=+173.505261045" watchObservedRunningTime="2026-03-21 00:14:20.213109218 +0000 UTC m=+173.506396400" Mar 21 00:14:20 crc kubenswrapper[5117]: I0321 00:14:20.271012 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:22 crc kubenswrapper[5117]: I0321 00:14:22.827171 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" podUID="7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" containerName="oauth-openshift" containerID="cri-o://4895ca8aa86f5389620bf2a7362ecb3f7ab72511f6c03a9de0137fccce21b86a" gracePeriod=15 Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.207167 5117 generic.go:358] "Generic (PLEG): container finished" podID="7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" containerID="4895ca8aa86f5389620bf2a7362ecb3f7ab72511f6c03a9de0137fccce21b86a" exitCode=0 Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.207331 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" event={"ID":"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2","Type":"ContainerDied","Data":"4895ca8aa86f5389620bf2a7362ecb3f7ab72511f6c03a9de0137fccce21b86a"} Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.293224 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.322388 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-869dfc464d-5662g"] Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.324073 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" containerName="oauth-openshift" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.324194 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" containerName="oauth-openshift" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.324585 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" containerName="oauth-openshift" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.334868 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-869dfc464d-5662g"] Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.335114 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.412776 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-provider-selection\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.412889 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-policies\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.412919 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-service-ca\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.412954 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-idp-0-file-data\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.412976 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-serving-cert\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413001 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-error\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413042 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-router-certs\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413063 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-dir\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413116 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb9gs\" (UniqueName: \"kubernetes.io/projected/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-kube-api-access-sb9gs\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413164 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-cliconfig\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413190 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-session\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413217 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-login\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413246 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-ocp-branding-template\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413302 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-trusted-ca-bundle\") pod \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\" (UID: \"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2\") " Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413426 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413459 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqlnp\" (UniqueName: \"kubernetes.io/projected/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-kube-api-access-hqlnp\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413492 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-session\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413534 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-template-login\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413558 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-audit-dir\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413578 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413598 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413640 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413676 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-service-ca\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413678 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.413700 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-router-certs\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.414080 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.414190 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.414615 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.415669 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.416000 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-audit-policies\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.416112 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.416160 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-template-error\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.416226 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.416304 5117 reconciler_common.go:299] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.416321 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.416332 5117 reconciler_common.go:299] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.416354 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.422485 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.422639 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.422826 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-kube-api-access-sb9gs" (OuterVolumeSpecName: "kube-api-access-sb9gs") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "kube-api-access-sb9gs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.423096 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.423315 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.423597 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.424056 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.424195 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.425640 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" (UID: "7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517447 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517499 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517530 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517554 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-service-ca\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517574 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-router-certs\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517601 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517620 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-audit-policies\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517639 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517659 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-template-error\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517696 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517717 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hqlnp\" (UniqueName: \"kubernetes.io/projected/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-kube-api-access-hqlnp\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517739 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-session\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517768 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-template-login\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517788 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-audit-dir\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517828 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517839 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517848 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517857 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517867 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-sb9gs\" (UniqueName: \"kubernetes.io/projected/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-kube-api-access-sb9gs\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517875 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517884 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517892 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517903 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517913 5117 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.517961 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-audit-dir\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.519131 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-audit-policies\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.521837 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-service-ca\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.522155 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.522238 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.522378 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.522406 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.523032 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.524427 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-template-error\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.524882 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.525066 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-user-template-login\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.525191 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-router-certs\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.526004 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-v4-0-config-system-session\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.542433 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqlnp\" (UniqueName: \"kubernetes.io/projected/a1f77a50-6ca5-4631-a632-a15c18b4f4b0-kube-api-access-hqlnp\") pod \"oauth-openshift-869dfc464d-5662g\" (UID: \"a1f77a50-6ca5-4631-a632-a15c18b4f4b0\") " pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.658488 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:23 crc kubenswrapper[5117]: I0321 00:14:23.838549 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-869dfc464d-5662g"] Mar 21 00:14:24 crc kubenswrapper[5117]: I0321 00:14:24.227163 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" Mar 21 00:14:24 crc kubenswrapper[5117]: I0321 00:14:24.227204 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-mtrnq" event={"ID":"7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2","Type":"ContainerDied","Data":"2db8c230993a06b01a23be9a47ff1dfdea780e559cef94464edba9e4a2f70d99"} Mar 21 00:14:24 crc kubenswrapper[5117]: I0321 00:14:24.227659 5117 scope.go:117] "RemoveContainer" containerID="4895ca8aa86f5389620bf2a7362ecb3f7ab72511f6c03a9de0137fccce21b86a" Mar 21 00:14:24 crc kubenswrapper[5117]: I0321 00:14:24.230170 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" event={"ID":"a1f77a50-6ca5-4631-a632-a15c18b4f4b0","Type":"ContainerStarted","Data":"a3e8e6ece73c0f157a72d0ec6429e78e53596f94580f1bb5f8fa033985179118"} Mar 21 00:14:24 crc kubenswrapper[5117]: I0321 00:14:24.248452 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mtrnq"] Mar 21 00:14:24 crc kubenswrapper[5117]: I0321 00:14:24.252041 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mtrnq"] Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.245883 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" event={"ID":"a1f77a50-6ca5-4631-a632-a15c18b4f4b0","Type":"ContainerStarted","Data":"00342029b7946219be59e5e433815dbe7157bc6ddbbaaa0cb8cbd5e7edae3844"} Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.246238 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.256129 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.275541 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-869dfc464d-5662g" podStartSLOduration=28.27551732 podStartE2EDuration="28.27551732s" podCreationTimestamp="2026-03-21 00:13:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:14:25.273974803 +0000 UTC m=+178.567262045" watchObservedRunningTime="2026-03-21 00:14:25.27551732 +0000 UTC m=+178.568804532" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.412670 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-12-crc"] Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.468861 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-12-crc"] Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.469040 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.472039 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver\"/\"kube-root-ca.crt\"" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.474022 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver\"/\"installer-sa-dockercfg-bqqnb\"" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.548478 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-var-lock\") pod \"installer-12-crc\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.548538 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kube-api-access\") pod \"installer-12-crc\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.548594 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kubelet-dir\") pod \"installer-12-crc\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.586866 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2" path="/var/lib/kubelet/pods/7aa6e0b1-e4e1-4f76-a7cd-8f9b7cba97d2/volumes" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.650643 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kube-api-access\") pod \"installer-12-crc\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.651089 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kubelet-dir\") pod \"installer-12-crc\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.651405 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kubelet-dir\") pod \"installer-12-crc\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.653235 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-var-lock\") pod \"installer-12-crc\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.653154 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-var-lock\") pod \"installer-12-crc\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.671020 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kube-api-access\") pod \"installer-12-crc\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:25 crc kubenswrapper[5117]: I0321 00:14:25.785931 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:14:26 crc kubenswrapper[5117]: I0321 00:14:26.171556 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-12-crc"] Mar 21 00:14:26 crc kubenswrapper[5117]: W0321 00:14:26.185215 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7739113f_88a1_4f1e_9fcf_d4045a611ed1.slice/crio-3b1e38c920c47d83c4e66b59ac8958ef7a48afb16d38d1ff94a35301b1662c4c WatchSource:0}: Error finding container 3b1e38c920c47d83c4e66b59ac8958ef7a48afb16d38d1ff94a35301b1662c4c: Status 404 returned error can't find the container with id 3b1e38c920c47d83c4e66b59ac8958ef7a48afb16d38d1ff94a35301b1662c4c Mar 21 00:14:26 crc kubenswrapper[5117]: I0321 00:14:26.258138 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"7739113f-88a1-4f1e-9fcf-d4045a611ed1","Type":"ContainerStarted","Data":"3b1e38c920c47d83c4e66b59ac8958ef7a48afb16d38d1ff94a35301b1662c4c"} Mar 21 00:14:27 crc kubenswrapper[5117]: I0321 00:14:27.267110 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"7739113f-88a1-4f1e-9fcf-d4045a611ed1","Type":"ContainerStarted","Data":"52881532ce3557e3edb1b199298d17e83b062fbdca97972715d09d23991b5694"} Mar 21 00:14:27 crc kubenswrapper[5117]: I0321 00:14:27.284398 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-12-crc" podStartSLOduration=2.2843817299999998 podStartE2EDuration="2.28438173s" podCreationTimestamp="2026-03-21 00:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:14:27.282662346 +0000 UTC m=+180.575949538" watchObservedRunningTime="2026-03-21 00:14:27.28438173 +0000 UTC m=+180.577668902" Mar 21 00:14:38 crc kubenswrapper[5117]: I0321 00:14:38.002993 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b468689cd-889sb"] Mar 21 00:14:38 crc kubenswrapper[5117]: I0321 00:14:38.005740 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" podUID="764d6012-5a93-4388-8efd-4a38a439ead8" containerName="controller-manager" containerID="cri-o://337223f3388f64fbb81d010acbf676df8616a2b4bd222dfcafa40ce38936c177" gracePeriod=30 Mar 21 00:14:38 crc kubenswrapper[5117]: I0321 00:14:38.018132 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj"] Mar 21 00:14:38 crc kubenswrapper[5117]: I0321 00:14:38.018789 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" podUID="76e297ff-c558-4d61-8d10-6ce8a183f450" containerName="route-controller-manager" containerID="cri-o://458e8653f23e665e6dd547636b7a0658a50275e8833ce2c79735b1f16f30b72d" gracePeriod=30 Mar 21 00:14:38 crc kubenswrapper[5117]: I0321 00:14:38.353500 5117 generic.go:358] "Generic (PLEG): container finished" podID="76e297ff-c558-4d61-8d10-6ce8a183f450" containerID="458e8653f23e665e6dd547636b7a0658a50275e8833ce2c79735b1f16f30b72d" exitCode=0 Mar 21 00:14:38 crc kubenswrapper[5117]: I0321 00:14:38.353573 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" event={"ID":"76e297ff-c558-4d61-8d10-6ce8a183f450","Type":"ContainerDied","Data":"458e8653f23e665e6dd547636b7a0658a50275e8833ce2c79735b1f16f30b72d"} Mar 21 00:14:38 crc kubenswrapper[5117]: I0321 00:14:38.355928 5117 generic.go:358] "Generic (PLEG): container finished" podID="764d6012-5a93-4388-8efd-4a38a439ead8" containerID="337223f3388f64fbb81d010acbf676df8616a2b4bd222dfcafa40ce38936c177" exitCode=0 Mar 21 00:14:38 crc kubenswrapper[5117]: I0321 00:14:38.356047 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" event={"ID":"764d6012-5a93-4388-8efd-4a38a439ead8","Type":"ContainerDied","Data":"337223f3388f64fbb81d010acbf676df8616a2b4bd222dfcafa40ce38936c177"} Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.156271 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.182876 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7"] Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.190707 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="76e297ff-c558-4d61-8d10-6ce8a183f450" containerName="route-controller-manager" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.190754 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e297ff-c558-4d61-8d10-6ce8a183f450" containerName="route-controller-manager" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.191130 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="76e297ff-c558-4d61-8d10-6ce8a183f450" containerName="route-controller-manager" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.293621 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-config\") pod \"76e297ff-c558-4d61-8d10-6ce8a183f450\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.293701 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76e297ff-c558-4d61-8d10-6ce8a183f450-serving-cert\") pod \"76e297ff-c558-4d61-8d10-6ce8a183f450\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.293743 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-client-ca\") pod \"76e297ff-c558-4d61-8d10-6ce8a183f450\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.293928 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxr67\" (UniqueName: \"kubernetes.io/projected/76e297ff-c558-4d61-8d10-6ce8a183f450-kube-api-access-pxr67\") pod \"76e297ff-c558-4d61-8d10-6ce8a183f450\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.294018 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/76e297ff-c558-4d61-8d10-6ce8a183f450-tmp\") pod \"76e297ff-c558-4d61-8d10-6ce8a183f450\" (UID: \"76e297ff-c558-4d61-8d10-6ce8a183f450\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.294295 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76e297ff-c558-4d61-8d10-6ce8a183f450-tmp" (OuterVolumeSpecName: "tmp") pod "76e297ff-c558-4d61-8d10-6ce8a183f450" (UID: "76e297ff-c558-4d61-8d10-6ce8a183f450"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.294631 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-client-ca" (OuterVolumeSpecName: "client-ca") pod "76e297ff-c558-4d61-8d10-6ce8a183f450" (UID: "76e297ff-c558-4d61-8d10-6ce8a183f450"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.294653 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-config" (OuterVolumeSpecName: "config") pod "76e297ff-c558-4d61-8d10-6ce8a183f450" (UID: "76e297ff-c558-4d61-8d10-6ce8a183f450"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.300722 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76e297ff-c558-4d61-8d10-6ce8a183f450-kube-api-access-pxr67" (OuterVolumeSpecName: "kube-api-access-pxr67") pod "76e297ff-c558-4d61-8d10-6ce8a183f450" (UID: "76e297ff-c558-4d61-8d10-6ce8a183f450"). InnerVolumeSpecName "kube-api-access-pxr67". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.302200 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76e297ff-c558-4d61-8d10-6ce8a183f450-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "76e297ff-c558-4d61-8d10-6ce8a183f450" (UID: "76e297ff-c558-4d61-8d10-6ce8a183f450"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.312286 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7"] Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.312453 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.363603 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" event={"ID":"76e297ff-c558-4d61-8d10-6ce8a183f450","Type":"ContainerDied","Data":"5ac5527c2a4aebf2addfba42598e9c48d1b3c101c5abea9f715bf73007dd1820"} Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.363664 5117 scope.go:117] "RemoveContainer" containerID="458e8653f23e665e6dd547636b7a0658a50275e8833ce2c79735b1f16f30b72d" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.363730 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.394941 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e900e07-b9fa-4462-9178-b360a5b8f789-config\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.395008 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e900e07-b9fa-4462-9178-b360a5b8f789-serving-cert\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.395093 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8e900e07-b9fa-4462-9178-b360a5b8f789-tmp\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.395226 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e900e07-b9fa-4462-9178-b360a5b8f789-client-ca\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.395481 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnjkg\" (UniqueName: \"kubernetes.io/projected/8e900e07-b9fa-4462-9178-b360a5b8f789-kube-api-access-fnjkg\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.395626 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.395648 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76e297ff-c558-4d61-8d10-6ce8a183f450-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.395662 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76e297ff-c558-4d61-8d10-6ce8a183f450-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.395676 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pxr67\" (UniqueName: \"kubernetes.io/projected/76e297ff-c558-4d61-8d10-6ce8a183f450-kube-api-access-pxr67\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.395688 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/76e297ff-c558-4d61-8d10-6ce8a183f450-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.400025 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj"] Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.405311 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54646bb8cc-259gj"] Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.430191 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.455865 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7856f74cf5-s5wcb"] Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.456547 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="764d6012-5a93-4388-8efd-4a38a439ead8" containerName="controller-manager" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.456570 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="764d6012-5a93-4388-8efd-4a38a439ead8" containerName="controller-manager" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.456703 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="764d6012-5a93-4388-8efd-4a38a439ead8" containerName="controller-manager" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496392 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/764d6012-5a93-4388-8efd-4a38a439ead8-tmp\") pod \"764d6012-5a93-4388-8efd-4a38a439ead8\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496478 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-proxy-ca-bundles\") pod \"764d6012-5a93-4388-8efd-4a38a439ead8\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496545 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxcwk\" (UniqueName: \"kubernetes.io/projected/764d6012-5a93-4388-8efd-4a38a439ead8-kube-api-access-kxcwk\") pod \"764d6012-5a93-4388-8efd-4a38a439ead8\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496616 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-config\") pod \"764d6012-5a93-4388-8efd-4a38a439ead8\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496650 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-client-ca\") pod \"764d6012-5a93-4388-8efd-4a38a439ead8\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496683 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/764d6012-5a93-4388-8efd-4a38a439ead8-serving-cert\") pod \"764d6012-5a93-4388-8efd-4a38a439ead8\" (UID: \"764d6012-5a93-4388-8efd-4a38a439ead8\") " Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496779 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e900e07-b9fa-4462-9178-b360a5b8f789-config\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496809 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e900e07-b9fa-4462-9178-b360a5b8f789-serving-cert\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496827 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8e900e07-b9fa-4462-9178-b360a5b8f789-tmp\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496859 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e900e07-b9fa-4462-9178-b360a5b8f789-client-ca\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.496921 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fnjkg\" (UniqueName: \"kubernetes.io/projected/8e900e07-b9fa-4462-9178-b360a5b8f789-kube-api-access-fnjkg\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.499063 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/764d6012-5a93-4388-8efd-4a38a439ead8-tmp" (OuterVolumeSpecName: "tmp") pod "764d6012-5a93-4388-8efd-4a38a439ead8" (UID: "764d6012-5a93-4388-8efd-4a38a439ead8"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.499286 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-client-ca" (OuterVolumeSpecName: "client-ca") pod "764d6012-5a93-4388-8efd-4a38a439ead8" (UID: "764d6012-5a93-4388-8efd-4a38a439ead8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.499591 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-config" (OuterVolumeSpecName: "config") pod "764d6012-5a93-4388-8efd-4a38a439ead8" (UID: "764d6012-5a93-4388-8efd-4a38a439ead8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.499643 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "764d6012-5a93-4388-8efd-4a38a439ead8" (UID: "764d6012-5a93-4388-8efd-4a38a439ead8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.500078 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8e900e07-b9fa-4462-9178-b360a5b8f789-tmp\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.500468 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e900e07-b9fa-4462-9178-b360a5b8f789-client-ca\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.501276 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e900e07-b9fa-4462-9178-b360a5b8f789-config\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.502039 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764d6012-5a93-4388-8efd-4a38a439ead8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "764d6012-5a93-4388-8efd-4a38a439ead8" (UID: "764d6012-5a93-4388-8efd-4a38a439ead8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.502616 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e900e07-b9fa-4462-9178-b360a5b8f789-serving-cert\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.503469 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/764d6012-5a93-4388-8efd-4a38a439ead8-kube-api-access-kxcwk" (OuterVolumeSpecName: "kube-api-access-kxcwk") pod "764d6012-5a93-4388-8efd-4a38a439ead8" (UID: "764d6012-5a93-4388-8efd-4a38a439ead8"). InnerVolumeSpecName "kube-api-access-kxcwk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.513785 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnjkg\" (UniqueName: \"kubernetes.io/projected/8e900e07-b9fa-4462-9178-b360a5b8f789-kube-api-access-fnjkg\") pod \"route-controller-manager-5478fcd6f5-dh4k7\" (UID: \"8e900e07-b9fa-4462-9178-b360a5b8f789\") " pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.537759 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7856f74cf5-s5wcb"] Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.537912 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.586873 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76e297ff-c558-4d61-8d10-6ce8a183f450" path="/var/lib/kubelet/pods/76e297ff-c558-4d61-8d10-6ce8a183f450/volumes" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598221 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19b66738-f1a1-47ae-a6a1-61c365a87dbc-proxy-ca-bundles\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598362 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/19b66738-f1a1-47ae-a6a1-61c365a87dbc-tmp\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598452 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krsqx\" (UniqueName: \"kubernetes.io/projected/19b66738-f1a1-47ae-a6a1-61c365a87dbc-kube-api-access-krsqx\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598526 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19b66738-f1a1-47ae-a6a1-61c365a87dbc-client-ca\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598585 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19b66738-f1a1-47ae-a6a1-61c365a87dbc-config\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598851 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19b66738-f1a1-47ae-a6a1-61c365a87dbc-serving-cert\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598941 5117 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-client-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598960 5117 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/764d6012-5a93-4388-8efd-4a38a439ead8-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598971 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/764d6012-5a93-4388-8efd-4a38a439ead8-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598982 5117 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.598994 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kxcwk\" (UniqueName: \"kubernetes.io/projected/764d6012-5a93-4388-8efd-4a38a439ead8-kube-api-access-kxcwk\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.599006 5117 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764d6012-5a93-4388-8efd-4a38a439ead8-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.646279 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.700640 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19b66738-f1a1-47ae-a6a1-61c365a87dbc-serving-cert\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.701193 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19b66738-f1a1-47ae-a6a1-61c365a87dbc-proxy-ca-bundles\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.701238 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/19b66738-f1a1-47ae-a6a1-61c365a87dbc-tmp\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.701328 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-krsqx\" (UniqueName: \"kubernetes.io/projected/19b66738-f1a1-47ae-a6a1-61c365a87dbc-kube-api-access-krsqx\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.703096 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19b66738-f1a1-47ae-a6a1-61c365a87dbc-client-ca\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.703900 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19b66738-f1a1-47ae-a6a1-61c365a87dbc-config\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.704031 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19b66738-f1a1-47ae-a6a1-61c365a87dbc-proxy-ca-bundles\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.704268 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/19b66738-f1a1-47ae-a6a1-61c365a87dbc-tmp\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.704725 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19b66738-f1a1-47ae-a6a1-61c365a87dbc-client-ca\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.705102 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19b66738-f1a1-47ae-a6a1-61c365a87dbc-config\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.708450 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19b66738-f1a1-47ae-a6a1-61c365a87dbc-serving-cert\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.733306 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-krsqx\" (UniqueName: \"kubernetes.io/projected/19b66738-f1a1-47ae-a6a1-61c365a87dbc-kube-api-access-krsqx\") pod \"controller-manager-7856f74cf5-s5wcb\" (UID: \"19b66738-f1a1-47ae-a6a1-61c365a87dbc\") " pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.852996 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:39 crc kubenswrapper[5117]: I0321 00:14:39.905488 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7"] Mar 21 00:14:39 crc kubenswrapper[5117]: W0321 00:14:39.914415 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e900e07_b9fa_4462_9178_b360a5b8f789.slice/crio-289206c0e4aa7a5883d583d609ed8f38d63b7fbbb69a1ac81fed58c73f1f89a0 WatchSource:0}: Error finding container 289206c0e4aa7a5883d583d609ed8f38d63b7fbbb69a1ac81fed58c73f1f89a0: Status 404 returned error can't find the container with id 289206c0e4aa7a5883d583d609ed8f38d63b7fbbb69a1ac81fed58c73f1f89a0 Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.068487 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7856f74cf5-s5wcb"] Mar 21 00:14:40 crc kubenswrapper[5117]: W0321 00:14:40.076074 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19b66738_f1a1_47ae_a6a1_61c365a87dbc.slice/crio-eb4e521355d418b7abd4279202a916b99e83172cdfaa9eea727da49f758d1485 WatchSource:0}: Error finding container eb4e521355d418b7abd4279202a916b99e83172cdfaa9eea727da49f758d1485: Status 404 returned error can't find the container with id eb4e521355d418b7abd4279202a916b99e83172cdfaa9eea727da49f758d1485 Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.370196 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" event={"ID":"19b66738-f1a1-47ae-a6a1-61c365a87dbc","Type":"ContainerStarted","Data":"f44212840bd61cdd1633940c315467b45b9d1c3aa5fd5e33c59c6808fa0d3c3b"} Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.370267 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" event={"ID":"19b66738-f1a1-47ae-a6a1-61c365a87dbc","Type":"ContainerStarted","Data":"eb4e521355d418b7abd4279202a916b99e83172cdfaa9eea727da49f758d1485"} Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.371569 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" event={"ID":"8e900e07-b9fa-4462-9178-b360a5b8f789","Type":"ContainerStarted","Data":"7e860265cc4a69923fe726c250629e1baac744aecc80323d830b962151b76bd1"} Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.371593 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" event={"ID":"8e900e07-b9fa-4462-9178-b360a5b8f789","Type":"ContainerStarted","Data":"289206c0e4aa7a5883d583d609ed8f38d63b7fbbb69a1ac81fed58c73f1f89a0"} Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.371798 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.378510 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" event={"ID":"764d6012-5a93-4388-8efd-4a38a439ead8","Type":"ContainerDied","Data":"3aed14f0b93b294998031674e374e96b0fe60212879d072412f9b8b9ba4cc7ea"} Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.378593 5117 scope.go:117] "RemoveContainer" containerID="337223f3388f64fbb81d010acbf676df8616a2b4bd222dfcafa40ce38936c177" Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.378648 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b468689cd-889sb" Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.395098 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" podStartSLOduration=2.395075711 podStartE2EDuration="2.395075711s" podCreationTimestamp="2026-03-21 00:14:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:14:40.391423668 +0000 UTC m=+193.684710840" watchObservedRunningTime="2026-03-21 00:14:40.395075711 +0000 UTC m=+193.688362893" Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.408883 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b468689cd-889sb"] Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.411216 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-b468689cd-889sb"] Mar 21 00:14:40 crc kubenswrapper[5117]: I0321 00:14:40.819518 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5478fcd6f5-dh4k7" Mar 21 00:14:41 crc kubenswrapper[5117]: I0321 00:14:41.391725 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:41 crc kubenswrapper[5117]: I0321 00:14:41.403798 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" Mar 21 00:14:41 crc kubenswrapper[5117]: I0321 00:14:41.433222 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7856f74cf5-s5wcb" podStartSLOduration=4.433133101 podStartE2EDuration="4.433133101s" podCreationTimestamp="2026-03-21 00:14:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:14:41.430784429 +0000 UTC m=+194.724071641" watchObservedRunningTime="2026-03-21 00:14:41.433133101 +0000 UTC m=+194.726420363" Mar 21 00:14:41 crc kubenswrapper[5117]: I0321 00:14:41.595616 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="764d6012-5a93-4388-8efd-4a38a439ead8" path="/var/lib/kubelet/pods/764d6012-5a93-4388-8efd-4a38a439ead8/volumes" Mar 21 00:14:45 crc kubenswrapper[5117]: I0321 00:14:45.436403 5117 generic.go:358] "Generic (PLEG): container finished" podID="79fcf4ae-c05f-492c-88f0-ec441eb255aa" containerID="ed9b114390e516515e6dce7ca532581b221ff2031f6a039a59d36381df3949f4" exitCode=0 Mar 21 00:14:45 crc kubenswrapper[5117]: I0321 00:14:45.436566 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29567520-5stkm" event={"ID":"79fcf4ae-c05f-492c-88f0-ec441eb255aa","Type":"ContainerDied","Data":"ed9b114390e516515e6dce7ca532581b221ff2031f6a039a59d36381df3949f4"} Mar 21 00:14:46 crc kubenswrapper[5117]: I0321 00:14:46.886466 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:14:46 crc kubenswrapper[5117]: I0321 00:14:46.936445 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/79fcf4ae-c05f-492c-88f0-ec441eb255aa-serviceca\") pod \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\" (UID: \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\") " Mar 21 00:14:46 crc kubenswrapper[5117]: I0321 00:14:46.936565 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkwnz\" (UniqueName: \"kubernetes.io/projected/79fcf4ae-c05f-492c-88f0-ec441eb255aa-kube-api-access-tkwnz\") pod \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\" (UID: \"79fcf4ae-c05f-492c-88f0-ec441eb255aa\") " Mar 21 00:14:46 crc kubenswrapper[5117]: I0321 00:14:46.938229 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79fcf4ae-c05f-492c-88f0-ec441eb255aa-serviceca" (OuterVolumeSpecName: "serviceca") pod "79fcf4ae-c05f-492c-88f0-ec441eb255aa" (UID: "79fcf4ae-c05f-492c-88f0-ec441eb255aa"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:14:46 crc kubenswrapper[5117]: I0321 00:14:46.944112 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79fcf4ae-c05f-492c-88f0-ec441eb255aa-kube-api-access-tkwnz" (OuterVolumeSpecName: "kube-api-access-tkwnz") pod "79fcf4ae-c05f-492c-88f0-ec441eb255aa" (UID: "79fcf4ae-c05f-492c-88f0-ec441eb255aa"). InnerVolumeSpecName "kube-api-access-tkwnz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:14:47 crc kubenswrapper[5117]: I0321 00:14:47.038340 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tkwnz\" (UniqueName: \"kubernetes.io/projected/79fcf4ae-c05f-492c-88f0-ec441eb255aa-kube-api-access-tkwnz\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:47 crc kubenswrapper[5117]: I0321 00:14:47.038397 5117 reconciler_common.go:299] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/79fcf4ae-c05f-492c-88f0-ec441eb255aa-serviceca\") on node \"crc\" DevicePath \"\"" Mar 21 00:14:47 crc kubenswrapper[5117]: I0321 00:14:47.458759 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29567520-5stkm" Mar 21 00:14:47 crc kubenswrapper[5117]: I0321 00:14:47.458742 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29567520-5stkm" event={"ID":"79fcf4ae-c05f-492c-88f0-ec441eb255aa","Type":"ContainerDied","Data":"f878db1c1959de3cbdb4308c50e29846a88745aecb3b9b05353134cbf045d89f"} Mar 21 00:14:47 crc kubenswrapper[5117]: I0321 00:14:47.458958 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f878db1c1959de3cbdb4308c50e29846a88745aecb3b9b05353134cbf045d89f" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.158574 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v"] Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.161095 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="79fcf4ae-c05f-492c-88f0-ec441eb255aa" containerName="image-pruner" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.161134 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="79fcf4ae-c05f-492c-88f0-ec441eb255aa" containerName="image-pruner" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.161482 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="79fcf4ae-c05f-492c-88f0-ec441eb255aa" containerName="image-pruner" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.175671 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v"] Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.175886 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.178738 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-dockercfg-vfqp6\"" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.178772 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-config\"" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.363442 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24442cf3-9234-456d-b719-80948679615f-config-volume\") pod \"collect-profiles-29567535-88w2v\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.363539 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24442cf3-9234-456d-b719-80948679615f-secret-volume\") pod \"collect-profiles-29567535-88w2v\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.363688 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s2gl\" (UniqueName: \"kubernetes.io/projected/24442cf3-9234-456d-b719-80948679615f-kube-api-access-7s2gl\") pod \"collect-profiles-29567535-88w2v\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.464757 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24442cf3-9234-456d-b719-80948679615f-config-volume\") pod \"collect-profiles-29567535-88w2v\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.464807 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24442cf3-9234-456d-b719-80948679615f-secret-volume\") pod \"collect-profiles-29567535-88w2v\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.464866 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7s2gl\" (UniqueName: \"kubernetes.io/projected/24442cf3-9234-456d-b719-80948679615f-kube-api-access-7s2gl\") pod \"collect-profiles-29567535-88w2v\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.466934 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24442cf3-9234-456d-b719-80948679615f-config-volume\") pod \"collect-profiles-29567535-88w2v\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.475755 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24442cf3-9234-456d-b719-80948679615f-secret-volume\") pod \"collect-profiles-29567535-88w2v\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.490752 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s2gl\" (UniqueName: \"kubernetes.io/projected/24442cf3-9234-456d-b719-80948679615f-kube-api-access-7s2gl\") pod \"collect-profiles-29567535-88w2v\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:00 crc kubenswrapper[5117]: I0321 00:15:00.499706 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:01 crc kubenswrapper[5117]: I0321 00:15:01.072390 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v"] Mar 21 00:15:01 crc kubenswrapper[5117]: W0321 00:15:01.082805 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24442cf3_9234_456d_b719_80948679615f.slice/crio-80e32545b1fe03172fce4cf22c0e4dbf6830870af02118132e34aaf739dbd2a4 WatchSource:0}: Error finding container 80e32545b1fe03172fce4cf22c0e4dbf6830870af02118132e34aaf739dbd2a4: Status 404 returned error can't find the container with id 80e32545b1fe03172fce4cf22c0e4dbf6830870af02118132e34aaf739dbd2a4 Mar 21 00:15:01 crc kubenswrapper[5117]: I0321 00:15:01.571194 5117 generic.go:358] "Generic (PLEG): container finished" podID="24442cf3-9234-456d-b719-80948679615f" containerID="8aa314cc6b98485eed51202cc44fc9b2c080d43fbf15f1477877b53619d9f114" exitCode=0 Mar 21 00:15:01 crc kubenswrapper[5117]: I0321 00:15:01.571293 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" event={"ID":"24442cf3-9234-456d-b719-80948679615f","Type":"ContainerDied","Data":"8aa314cc6b98485eed51202cc44fc9b2c080d43fbf15f1477877b53619d9f114"} Mar 21 00:15:01 crc kubenswrapper[5117]: I0321 00:15:01.571711 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" event={"ID":"24442cf3-9234-456d-b719-80948679615f","Type":"ContainerStarted","Data":"80e32545b1fe03172fce4cf22c0e4dbf6830870af02118132e34aaf739dbd2a4"} Mar 21 00:15:02 crc kubenswrapper[5117]: I0321 00:15:02.999517 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.101567 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24442cf3-9234-456d-b719-80948679615f-config-volume\") pod \"24442cf3-9234-456d-b719-80948679615f\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.101661 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24442cf3-9234-456d-b719-80948679615f-secret-volume\") pod \"24442cf3-9234-456d-b719-80948679615f\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.101772 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s2gl\" (UniqueName: \"kubernetes.io/projected/24442cf3-9234-456d-b719-80948679615f-kube-api-access-7s2gl\") pod \"24442cf3-9234-456d-b719-80948679615f\" (UID: \"24442cf3-9234-456d-b719-80948679615f\") " Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.104159 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24442cf3-9234-456d-b719-80948679615f-config-volume" (OuterVolumeSpecName: "config-volume") pod "24442cf3-9234-456d-b719-80948679615f" (UID: "24442cf3-9234-456d-b719-80948679615f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.112792 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24442cf3-9234-456d-b719-80948679615f-kube-api-access-7s2gl" (OuterVolumeSpecName: "kube-api-access-7s2gl") pod "24442cf3-9234-456d-b719-80948679615f" (UID: "24442cf3-9234-456d-b719-80948679615f"). InnerVolumeSpecName "kube-api-access-7s2gl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.113393 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24442cf3-9234-456d-b719-80948679615f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "24442cf3-9234-456d-b719-80948679615f" (UID: "24442cf3-9234-456d-b719-80948679615f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.204669 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7s2gl\" (UniqueName: \"kubernetes.io/projected/24442cf3-9234-456d-b719-80948679615f-kube-api-access-7s2gl\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.204736 5117 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24442cf3-9234-456d-b719-80948679615f-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.204758 5117 reconciler_common.go:299] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24442cf3-9234-456d-b719-80948679615f-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.588475 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.592018 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567535-88w2v" event={"ID":"24442cf3-9234-456d-b719-80948679615f","Type":"ContainerDied","Data":"80e32545b1fe03172fce4cf22c0e4dbf6830870af02118132e34aaf739dbd2a4"} Mar 21 00:15:03 crc kubenswrapper[5117]: I0321 00:15:03.592087 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80e32545b1fe03172fce4cf22c0e4dbf6830870af02118132e34aaf739dbd2a4" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.549024 5117 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.550119 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="24442cf3-9234-456d-b719-80948679615f" containerName="collect-profiles" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.550147 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="24442cf3-9234-456d-b719-80948679615f" containerName="collect-profiles" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.550314 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="24442cf3-9234-456d-b719-80948679615f" containerName="collect-profiles" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.613698 5117 kubelet.go:2547] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.613850 5117 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.614105 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615217 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615472 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615507 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615522 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615560 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615577 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615606 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="setup" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615618 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="setup" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615634 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615647 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615660 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615672 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615694 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615707 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615735 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615751 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615732 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" containerID="cri-o://31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374" gracePeriod=15 Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615770 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" containerID="cri-o://460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd" gracePeriod=15 Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615804 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" containerID="cri-o://41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae" gracePeriod=15 Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615767 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61" gracePeriod=15 Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615780 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615995 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.615943 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7" gracePeriod=15 Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.618974 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619016 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619034 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619058 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619082 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619098 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619122 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619138 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619391 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619412 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.619667 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.626932 5117 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="3a14caf222afb62aaabdc47808b6f944" podUID="57755cc5f99000cc11e193051474d4e2" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.666052 5117 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730092 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730242 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730359 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730392 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730437 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730483 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730527 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730565 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730594 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.730624 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832635 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832701 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832733 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832757 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832765 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832823 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832826 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832857 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832895 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832898 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832938 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832963 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832968 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.833001 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.833004 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.833035 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.833142 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.832938 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.833341 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:04 crc kubenswrapper[5117]: I0321 00:15:04.833405 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.610296 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.613501 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.614952 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374" exitCode=0 Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.615019 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7" exitCode=0 Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.615043 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61" exitCode=0 Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.615062 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae" exitCode=2 Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.615412 5117 scope.go:117] "RemoveContainer" containerID="79c4c26191e55d73237364c4a2aa0a9f41e94a520d1d505c560970920ec08d17" Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.625639 5117 generic.go:358] "Generic (PLEG): container finished" podID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" containerID="52881532ce3557e3edb1b199298d17e83b062fbdca97972715d09d23991b5694" exitCode=0 Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.625759 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"7739113f-88a1-4f1e-9fcf-d4045a611ed1","Type":"ContainerDied","Data":"52881532ce3557e3edb1b199298d17e83b062fbdca97972715d09d23991b5694"} Mar 21 00:15:05 crc kubenswrapper[5117]: I0321 00:15:05.627448 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:06 crc kubenswrapper[5117]: I0321 00:15:06.639048 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.127727 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.129289 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.172538 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kube-api-access\") pod \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.172678 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kubelet-dir\") pod \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.172846 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-var-lock\") pod \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\" (UID: \"7739113f-88a1-4f1e-9fcf-d4045a611ed1\") " Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.172956 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7739113f-88a1-4f1e-9fcf-d4045a611ed1" (UID: "7739113f-88a1-4f1e-9fcf-d4045a611ed1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.173069 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-var-lock" (OuterVolumeSpecName: "var-lock") pod "7739113f-88a1-4f1e-9fcf-d4045a611ed1" (UID: "7739113f-88a1-4f1e-9fcf-d4045a611ed1"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.173345 5117 reconciler_common.go:299] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-var-lock\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.173391 5117 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.181901 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7739113f-88a1-4f1e-9fcf-d4045a611ed1" (UID: "7739113f-88a1-4f1e-9fcf-d4045a611ed1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.274958 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7739113f-88a1-4f1e-9fcf-d4045a611ed1-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.516640 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.517897 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.518775 5117 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.519203 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.580477 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.581066 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.581156 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.581478 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.581552 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.581799 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.581945 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.581864 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.582442 5117 reconciler_common.go:299] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.582570 5117 reconciler_common.go:299] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.582684 5117 reconciler_common.go:299] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.582714 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir" (OuterVolumeSpecName: "ca-bundle-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "ca-bundle-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.585632 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.585827 5117 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.586397 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.594457 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a14caf222afb62aaabdc47808b6f944" path="/var/lib/kubelet/pods/3a14caf222afb62aaabdc47808b6f944/volumes" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.652206 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.652989 5117 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd" exitCode=0 Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.653425 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.653529 5117 scope.go:117] "RemoveContainer" containerID="31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.654195 5117 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.654717 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.656869 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"7739113f-88a1-4f1e-9fcf-d4045a611ed1","Type":"ContainerDied","Data":"3b1e38c920c47d83c4e66b59ac8958ef7a48afb16d38d1ff94a35301b1662c4c"} Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.656910 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b1e38c920c47d83c4e66b59ac8958ef7a48afb16d38d1ff94a35301b1662c4c" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.656995 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.658665 5117 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.659632 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.661772 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.662511 5117 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.675487 5117 scope.go:117] "RemoveContainer" containerID="64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.684058 5117 reconciler_common.go:299] "Volume detached for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.684172 5117 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.710086 5117 scope.go:117] "RemoveContainer" containerID="4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.729670 5117 scope.go:117] "RemoveContainer" containerID="41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.754598 5117 scope.go:117] "RemoveContainer" containerID="460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.775373 5117 scope.go:117] "RemoveContainer" containerID="dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.852415 5117 scope.go:117] "RemoveContainer" containerID="31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374" Mar 21 00:15:07 crc kubenswrapper[5117]: E0321 00:15:07.852993 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374\": container with ID starting with 31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374 not found: ID does not exist" containerID="31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.853121 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374"} err="failed to get container status \"31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374\": rpc error: code = NotFound desc = could not find container \"31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374\": container with ID starting with 31561dba23ee3c778efc6c271817d0ac54b9e5befde856d801d2f4a4272c6374 not found: ID does not exist" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.853157 5117 scope.go:117] "RemoveContainer" containerID="64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7" Mar 21 00:15:07 crc kubenswrapper[5117]: E0321 00:15:07.853427 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7\": container with ID starting with 64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7 not found: ID does not exist" containerID="64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.853507 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7"} err="failed to get container status \"64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7\": rpc error: code = NotFound desc = could not find container \"64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7\": container with ID starting with 64899ab86ba775fc698ff0ea4bd1d017565bb6bb0620baba350366f1ad447af7 not found: ID does not exist" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.853528 5117 scope.go:117] "RemoveContainer" containerID="4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61" Mar 21 00:15:07 crc kubenswrapper[5117]: E0321 00:15:07.853725 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61\": container with ID starting with 4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61 not found: ID does not exist" containerID="4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.853757 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61"} err="failed to get container status \"4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61\": rpc error: code = NotFound desc = could not find container \"4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61\": container with ID starting with 4302bad595ef86c934810d8fa534c426f08ab81069cb48feb2aba1dc443c5b61 not found: ID does not exist" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.853840 5117 scope.go:117] "RemoveContainer" containerID="41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae" Mar 21 00:15:07 crc kubenswrapper[5117]: E0321 00:15:07.854421 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae\": container with ID starting with 41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae not found: ID does not exist" containerID="41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.854457 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae"} err="failed to get container status \"41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae\": rpc error: code = NotFound desc = could not find container \"41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae\": container with ID starting with 41d1946b855a9a0c568e62ebfca2459841209f433851bae641c661c7a079daae not found: ID does not exist" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.854482 5117 scope.go:117] "RemoveContainer" containerID="460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd" Mar 21 00:15:07 crc kubenswrapper[5117]: E0321 00:15:07.854737 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd\": container with ID starting with 460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd not found: ID does not exist" containerID="460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.854779 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd"} err="failed to get container status \"460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd\": rpc error: code = NotFound desc = could not find container \"460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd\": container with ID starting with 460b84e2274fc9f82589b8e0037efcd62da9b118ac4227c155210b9fed54cbfd not found: ID does not exist" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.854802 5117 scope.go:117] "RemoveContainer" containerID="dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a" Mar 21 00:15:07 crc kubenswrapper[5117]: E0321 00:15:07.855021 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a\": container with ID starting with dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a not found: ID does not exist" containerID="dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a" Mar 21 00:15:07 crc kubenswrapper[5117]: I0321 00:15:07.855054 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a"} err="failed to get container status \"dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a\": rpc error: code = NotFound desc = could not find container \"dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a\": container with ID starting with dd3d892e2464c919a3a5911c0dd88eb9d9cdd7aae512cedfac4c38b0c4e4eb1a not found: ID does not exist" Mar 21 00:15:09 crc kubenswrapper[5117]: E0321 00:15:09.308653 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:15:09Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:15:09Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:15:09Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-21T00:15:09Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:09 crc kubenswrapper[5117]: E0321 00:15:09.310687 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:09 crc kubenswrapper[5117]: E0321 00:15:09.311127 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:09 crc kubenswrapper[5117]: E0321 00:15:09.311649 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:09 crc kubenswrapper[5117]: E0321 00:15:09.312422 5117 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:09 crc kubenswrapper[5117]: E0321 00:15:09.312504 5117 kubelet_node_status.go:584] "Unable to update node status" err="update node status exceeds retry count" Mar 21 00:15:09 crc kubenswrapper[5117]: E0321 00:15:09.668914 5117 kubelet.go:3342] "Failed creating a mirror pod" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.243:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:09 crc kubenswrapper[5117]: I0321 00:15:09.670161 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:09 crc kubenswrapper[5117]: E0321 00:15:09.670947 5117 desired_state_of_world_populator.go:305] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.243:6443: connect: connection refused" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" volumeName="registry-storage" Mar 21 00:15:09 crc kubenswrapper[5117]: E0321 00:15:09.711413 5117 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.243:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189eb2fe942e1ab6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f7dbc7e1ee9c187a863ef9b473fad27b,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:15:09.708040886 +0000 UTC m=+223.001328098,LastTimestamp:2026-03-21 00:15:09.708040886 +0000 UTC m=+223.001328098,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:15:10 crc kubenswrapper[5117]: I0321 00:15:10.685304 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f7dbc7e1ee9c187a863ef9b473fad27b","Type":"ContainerStarted","Data":"5ded4898674941b2ee00c3899106d71bb52f10164dc020f73a0c47e05171f1cb"} Mar 21 00:15:10 crc kubenswrapper[5117]: I0321 00:15:10.687346 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f7dbc7e1ee9c187a863ef9b473fad27b","Type":"ContainerStarted","Data":"c34875cbcb668d66fad6cff775a1cee539b0df1ff1d577d2d86a83805a638370"} Mar 21 00:15:10 crc kubenswrapper[5117]: I0321 00:15:10.687801 5117 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:10 crc kubenswrapper[5117]: E0321 00:15:10.688503 5117 kubelet.go:3342] "Failed creating a mirror pod" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.243:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:10 crc kubenswrapper[5117]: I0321 00:15:10.688560 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:11 crc kubenswrapper[5117]: E0321 00:15:11.913757 5117 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:11 crc kubenswrapper[5117]: E0321 00:15:11.914765 5117 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:11 crc kubenswrapper[5117]: E0321 00:15:11.915440 5117 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:11 crc kubenswrapper[5117]: E0321 00:15:11.916172 5117 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:11 crc kubenswrapper[5117]: E0321 00:15:11.917364 5117 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:11 crc kubenswrapper[5117]: I0321 00:15:11.917430 5117 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 21 00:15:11 crc kubenswrapper[5117]: E0321 00:15:11.917988 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="200ms" Mar 21 00:15:12 crc kubenswrapper[5117]: E0321 00:15:12.119427 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="400ms" Mar 21 00:15:12 crc kubenswrapper[5117]: E0321 00:15:12.521124 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="800ms" Mar 21 00:15:12 crc kubenswrapper[5117]: E0321 00:15:12.723661 5117 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.243:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189eb2fe942e1ab6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f7dbc7e1ee9c187a863ef9b473fad27b,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-21 00:15:09.708040886 +0000 UTC m=+223.001328098,LastTimestamp:2026-03-21 00:15:09.708040886 +0000 UTC m=+223.001328098,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 21 00:15:13 crc kubenswrapper[5117]: E0321 00:15:13.323390 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="1.6s" Mar 21 00:15:14 crc kubenswrapper[5117]: E0321 00:15:14.924503 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="3.2s" Mar 21 00:15:17 crc kubenswrapper[5117]: I0321 00:15:17.587383 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:17 crc kubenswrapper[5117]: I0321 00:15:17.587863 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:17 crc kubenswrapper[5117]: I0321 00:15:17.592691 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:17 crc kubenswrapper[5117]: I0321 00:15:17.620746 5117 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:17 crc kubenswrapper[5117]: I0321 00:15:17.620810 5117 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:17 crc kubenswrapper[5117]: E0321 00:15:17.621726 5117 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:17 crc kubenswrapper[5117]: I0321 00:15:17.622204 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:17 crc kubenswrapper[5117]: I0321 00:15:17.752068 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"b9c7f4591431696d18076c308124160b8787f78d3e54c8ac3aae084123504de8"} Mar 21 00:15:18 crc kubenswrapper[5117]: E0321 00:15:18.126996 5117 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.243:6443: connect: connection refused" interval="6.4s" Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.761588 5117 generic.go:358] "Generic (PLEG): container finished" podID="57755cc5f99000cc11e193051474d4e2" containerID="dca9f235fb8fc183292c01ff20489c25e1a9ab67d960f2fc2b9294ca9b8c3be5" exitCode=0 Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.761673 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerDied","Data":"dca9f235fb8fc183292c01ff20489c25e1a9ab67d960f2fc2b9294ca9b8c3be5"} Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.762301 5117 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.762347 5117 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.762803 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:18 crc kubenswrapper[5117]: E0321 00:15:18.762835 5117 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.767082 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.767143 5117 generic.go:358] "Generic (PLEG): container finished" podID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerID="ad24a3d986e06d5c3f1905aa7e44d8fcc888cb7acc101ca3cbb39adac8684500" exitCode=1 Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.767247 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerDied","Data":"ad24a3d986e06d5c3f1905aa7e44d8fcc888cb7acc101ca3cbb39adac8684500"} Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.767855 5117 scope.go:117] "RemoveContainer" containerID="ad24a3d986e06d5c3f1905aa7e44d8fcc888cb7acc101ca3cbb39adac8684500" Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.768433 5117 status_manager.go:895] "Failed to get status for pod" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.769097 5117 status_manager.go:895] "Failed to get status for pod" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.243:6443: connect: connection refused" Mar 21 00:15:18 crc kubenswrapper[5117]: I0321 00:15:18.843503 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:15:19 crc kubenswrapper[5117]: I0321 00:15:19.788310 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"7cd238b78dc92137245a14a4e99797f59078fe65a31a4063e078d1fcc0c321c9"} Mar 21 00:15:19 crc kubenswrapper[5117]: I0321 00:15:19.788863 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"a4ff12d801e3117a1452ae040ec3227277486cfd2095ee47034f46156086682d"} Mar 21 00:15:19 crc kubenswrapper[5117]: I0321 00:15:19.792586 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:15:19 crc kubenswrapper[5117]: I0321 00:15:19.792748 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"c092a3ac53767d81078536eb992bc0a9fa364eade0ca87b7b4df05bd1ff144af"} Mar 21 00:15:20 crc kubenswrapper[5117]: I0321 00:15:20.807032 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"599a366aa544ad1d24f2b752069b03cdf0727ec6450e4b67a622f4eaab0e006b"} Mar 21 00:15:20 crc kubenswrapper[5117]: I0321 00:15:20.807658 5117 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:20 crc kubenswrapper[5117]: I0321 00:15:20.807704 5117 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:20 crc kubenswrapper[5117]: I0321 00:15:20.807826 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:20 crc kubenswrapper[5117]: I0321 00:15:20.807851 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"337212c8e2bd99b5649d8437f2ff6b30ed6e3437488e2d9eac40c80bbc639428"} Mar 21 00:15:20 crc kubenswrapper[5117]: I0321 00:15:20.807864 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"c17a8fcad1cb7ef1177cffd9fe1b0346cb5732817b7817c64d5d21e2c3b9b989"} Mar 21 00:15:22 crc kubenswrapper[5117]: I0321 00:15:22.622738 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:22 crc kubenswrapper[5117]: I0321 00:15:22.623307 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:22 crc kubenswrapper[5117]: I0321 00:15:22.629711 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:24 crc kubenswrapper[5117]: I0321 00:15:24.718720 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:15:24 crc kubenswrapper[5117]: I0321 00:15:24.718862 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:15:25 crc kubenswrapper[5117]: I0321 00:15:25.822580 5117 kubelet.go:3329] "Deleted mirror pod as it didn't match the static Pod" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:25 crc kubenswrapper[5117]: I0321 00:15:25.822896 5117 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:25 crc kubenswrapper[5117]: I0321 00:15:25.864006 5117 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:25 crc kubenswrapper[5117]: I0321 00:15:25.864040 5117 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:25 crc kubenswrapper[5117]: I0321 00:15:25.871986 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:26 crc kubenswrapper[5117]: I0321 00:15:26.894037 5117 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:26 crc kubenswrapper[5117]: I0321 00:15:26.894070 5117 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="797d5d7c-c536-4487-99be-f1f0219b63c6" Mar 21 00:15:27 crc kubenswrapper[5117]: I0321 00:15:27.603391 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:15:27 crc kubenswrapper[5117]: I0321 00:15:27.609756 5117 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="57755cc5f99000cc11e193051474d4e2" podUID="df725843-a674-4566-bac5-51435efe35c1" Mar 21 00:15:27 crc kubenswrapper[5117]: I0321 00:15:27.609904 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:15:27 crc kubenswrapper[5117]: I0321 00:15:27.674036 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:15:36 crc kubenswrapper[5117]: I0321 00:15:36.296169 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:36 crc kubenswrapper[5117]: I0321 00:15:36.872660 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Mar 21 00:15:37 crc kubenswrapper[5117]: I0321 00:15:37.484764 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-operator-tls\"" Mar 21 00:15:37 crc kubenswrapper[5117]: I0321 00:15:37.504164 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-global-ca\"" Mar 21 00:15:37 crc kubenswrapper[5117]: I0321 00:15:37.509120 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"kube-root-ca.crt\"" Mar 21 00:15:37 crc kubenswrapper[5117]: I0321 00:15:37.689170 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"env-overrides\"" Mar 21 00:15:37 crc kubenswrapper[5117]: I0321 00:15:37.842800 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:37 crc kubenswrapper[5117]: I0321 00:15:37.890880 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-6tbpn\"" Mar 21 00:15:38 crc kubenswrapper[5117]: I0321 00:15:38.210132 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Mar 21 00:15:38 crc kubenswrapper[5117]: I0321 00:15:38.620846 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:38 crc kubenswrapper[5117]: I0321 00:15:38.741062 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:38 crc kubenswrapper[5117]: I0321 00:15:38.918133 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 21 00:15:38 crc kubenswrapper[5117]: I0321 00:15:38.918166 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"default-dockercfg-hqpm5\"" Mar 21 00:15:38 crc kubenswrapper[5117]: I0321 00:15:38.961374 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"openshift-service-ca.crt\"" Mar 21 00:15:38 crc kubenswrapper[5117]: I0321 00:15:38.962636 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-serving-cert\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.019610 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-jmhxf\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.034213 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"kube-root-ca.crt\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.069392 5117 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.162680 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-dockercfg-sw6nc\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.171328 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"openshift-config-operator-dockercfg-sjn6s\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.273856 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.444685 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-serving-cert\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.485293 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.491287 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.559469 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-images\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.590752 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-rbac-proxy\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.631065 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-root-ca.crt\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.638659 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.739353 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"client-ca\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.793283 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-root-ca.crt\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.876412 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.931680 5117 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.940173 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.940297 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.947739 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 21 00:15:39 crc kubenswrapper[5117]: I0321 00:15:39.980884 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=14.980858182 podStartE2EDuration="14.980858182s" podCreationTimestamp="2026-03-21 00:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:15:39.973487512 +0000 UTC m=+253.266774684" watchObservedRunningTime="2026-03-21 00:15:39.980858182 +0000 UTC m=+253.274145364" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.213224 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.247624 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.265616 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.309613 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.314374 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.323606 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-metrics-certs-default\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.387830 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.453951 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.468373 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.477963 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.480750 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.480758 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-session\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.498497 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.525134 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-config\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.609642 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.657479 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-bgxvm\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.741801 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"image-import-ca\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.799101 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.838957 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-t8n29\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.843743 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"serving-cert\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.845715 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.857298 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"cluster-image-registry-operator-dockercfg-ntnd7\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.869568 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-mmcpt\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.876610 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.911140 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\"" Mar 21 00:15:40 crc kubenswrapper[5117]: I0321 00:15:40.981397 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-images\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.013867 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-qqw4z\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.063451 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.146005 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.151684 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.164214 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"ingress-operator-dockercfg-74nwh\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.186193 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-dockercfg-2cfkp\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.240457 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-ca-bundle\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.261413 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-service-ca-bundle\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.344968 5117 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.356334 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.616454 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-metrics\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.693424 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.762873 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.818892 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-certs-default\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.824108 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"etcd-client\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.833433 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-client\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.917878 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-dockercfg-gnx66\"" Mar 21 00:15:41 crc kubenswrapper[5117]: I0321 00:15:41.971953 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"networking-console-plugin-cert\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.037562 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-dockercfg-6c46w\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.110987 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"audit-1\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.112167 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-7cl8d\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.234744 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-kw8fx\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.288961 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-9gxlh\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.372808 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.377494 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-dockercfg-jcmfj\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.435246 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-console\"/\"networking-console-plugin\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.463719 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.494902 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.520808 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-nwglk\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.522709 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-djmfg\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.615726 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.644744 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.662957 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-tls\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.762310 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-dockercfg-tnfx9\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.839433 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"openshift-service-ca.crt\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.846545 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-scheduler-operator-serving-cert\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.847690 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Mar 21 00:15:42 crc kubenswrapper[5117]: I0321 00:15:42.904628 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"serving-cert\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.086417 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.095035 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"community-operators-dockercfg-vrd5f\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.126491 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"service-ca-bundle\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.182333 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.252916 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-dockercfg-4vdnc\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.304350 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"openshift-service-ca.crt\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.305899 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-tls\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.358076 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"kube-root-ca.crt\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.363068 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.396928 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"serving-cert\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.403687 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mco-proxy-tls\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.427399 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.529923 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-l2v2m\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.530817 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-6w67b\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.552654 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"ovnkube-identity-cm\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.704494 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-tls\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.709001 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"kube-root-ca.crt\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.763454 5117 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.770779 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"pprof-cert\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.854408 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:43 crc kubenswrapper[5117]: I0321 00:15:43.942660 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.087113 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-kknhg\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.119964 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.144873 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.146923 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"metrics-tls\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.169365 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"packageserver-service-cert\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.267808 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-config\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.268644 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-serving-cert\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.317058 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-dockercfg-dzw6b\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.381846 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-config\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.384365 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-dockercfg-2wbn2\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.535533 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-4zqgh\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.592190 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"openshift-service-ca.crt\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.594404 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.727350 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.774785 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"config\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.855718 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.881654 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"trusted-ca-bundle\"" Mar 21 00:15:44 crc kubenswrapper[5117]: I0321 00:15:44.969401 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-kpvmz\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.004459 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.040179 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.043174 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mcc-proxy-tls\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.096049 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-serving-cert\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.132413 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.205453 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-serving-cert\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.215217 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"hostpath-provisioner\"/\"csi-hostpath-provisioner-sa-dockercfg-7dcws\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.240733 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-gg4w7\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.241442 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.259367 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"kube-root-ca.crt\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.262243 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-config\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.444338 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.500802 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"node-bootstrapper-token\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.503419 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.567785 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"metrics-tls\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.653933 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.696906 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"config\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.751736 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"service-ca-bundle\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.756336 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-g6kgg\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.790227 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.881067 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.941024 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.968581 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"catalog-operator-serving-cert\"" Mar 21 00:15:45 crc kubenswrapper[5117]: I0321 00:15:45.982565 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-router-certs\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.012012 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-dockercfg-bjqfd\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.059084 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-config\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.186440 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.229814 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-stats-default\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.244483 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-admission-controller-secret\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.284189 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"serving-cert\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.294452 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-wzhvk\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.301787 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-dockercfg-bf7fj\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.329672 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"trusted-ca-bundle\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.346117 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-node-identity\"/\"network-node-identity-cert\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.359002 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.396109 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.488840 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.547057 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.621713 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.624115 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"etcd-serving-ca\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.645183 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"cluster-version-operator-serving-cert\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.656976 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.752716 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.772955 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"client-ca\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.815862 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.816059 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.821668 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Mar 21 00:15:46 crc kubenswrapper[5117]: I0321 00:15:46.944786 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"serving-cert\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.039167 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.148951 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.152926 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-config\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.287561 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.288128 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.374427 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"dns-operator-dockercfg-wbbsn\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.438026 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-ocp-branding-template\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.555188 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-controller-dockercfg-xnj77\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.592228 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.611104 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.691450 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-client\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.734953 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.802198 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.920415 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:47 crc kubenswrapper[5117]: I0321 00:15:47.951086 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.029283 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-tjs74\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.081110 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"config\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.147370 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.150229 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-6n5ln\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.371917 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.375143 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.403981 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"openshift-service-ca.crt\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.484536 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.514632 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.515821 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.518526 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-w9nzh\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.558056 5117 kubelet.go:2547] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.558913 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" containerID="cri-o://5ded4898674941b2ee00c3899106d71bb52f10164dc020f73a0c47e05171f1cb" gracePeriod=5 Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.731207 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9pgs7\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.732238 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.788150 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serviceaccount-dockercfg-4gqzj\"" Mar 21 00:15:48 crc kubenswrapper[5117]: I0321 00:15:48.988785 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"serving-cert\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.062897 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-tk7bt\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.079608 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.097729 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"encryption-config-1\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.116364 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"audit\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.137936 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-rbac-proxy\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.204428 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.219089 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-kl6m8\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.230888 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-8dkm8\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.250289 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ac-dockercfg-gj7jx\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.262481 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.441205 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.569275 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-config\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.659269 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-config\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.715725 5117 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.803407 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Mar 21 00:15:49 crc kubenswrapper[5117]: I0321 00:15:49.965868 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"kube-root-ca.crt\"" Mar 21 00:15:50 crc kubenswrapper[5117]: I0321 00:15:50.037124 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-d2bf2\"" Mar 21 00:15:50 crc kubenswrapper[5117]: I0321 00:15:50.082188 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Mar 21 00:15:50 crc kubenswrapper[5117]: I0321 00:15:50.261417 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:50 crc kubenswrapper[5117]: I0321 00:15:50.285275 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\"" Mar 21 00:15:50 crc kubenswrapper[5117]: I0321 00:15:50.412746 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"config-operator-serving-cert\"" Mar 21 00:15:50 crc kubenswrapper[5117]: I0321 00:15:50.456475 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\"" Mar 21 00:15:50 crc kubenswrapper[5117]: I0321 00:15:50.499514 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Mar 21 00:15:50 crc kubenswrapper[5117]: I0321 00:15:50.559386 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"metrics-tls\"" Mar 21 00:15:50 crc kubenswrapper[5117]: I0321 00:15:50.927633 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-config\"" Mar 21 00:15:51 crc kubenswrapper[5117]: I0321 00:15:51.126899 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Mar 21 00:15:51 crc kubenswrapper[5117]: I0321 00:15:51.255659 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Mar 21 00:15:51 crc kubenswrapper[5117]: I0321 00:15:51.299656 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serving-cert\"" Mar 21 00:15:51 crc kubenswrapper[5117]: I0321 00:15:51.349418 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"marketplace-trusted-ca\"" Mar 21 00:15:51 crc kubenswrapper[5117]: I0321 00:15:51.476763 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Mar 21 00:15:51 crc kubenswrapper[5117]: I0321 00:15:51.550123 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"default-dockercfg-mdwwj\"" Mar 21 00:15:51 crc kubenswrapper[5117]: I0321 00:15:51.580047 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Mar 21 00:15:51 crc kubenswrapper[5117]: I0321 00:15:51.721824 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-tls\"" Mar 21 00:15:51 crc kubenswrapper[5117]: I0321 00:15:51.800717 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Mar 21 00:15:52 crc kubenswrapper[5117]: I0321 00:15:52.084084 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"audit-1\"" Mar 21 00:15:52 crc kubenswrapper[5117]: I0321 00:15:52.383646 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"package-server-manager-serving-cert\"" Mar 21 00:15:52 crc kubenswrapper[5117]: I0321 00:15:52.446236 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"trusted-ca\"" Mar 21 00:15:52 crc kubenswrapper[5117]: I0321 00:15:52.615308 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-nl8tp\"" Mar 21 00:15:52 crc kubenswrapper[5117]: I0321 00:15:52.832701 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"proxy-tls\"" Mar 21 00:15:53 crc kubenswrapper[5117]: I0321 00:15:53.363328 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"kube-root-ca.crt\"" Mar 21 00:15:53 crc kubenswrapper[5117]: I0321 00:15:53.512411 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-2h6bs\"" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.132476 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f7dbc7e1ee9c187a863ef9b473fad27b/startup-monitor/0.log" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.132979 5117 generic.go:358] "Generic (PLEG): container finished" podID="f7dbc7e1ee9c187a863ef9b473fad27b" containerID="5ded4898674941b2ee00c3899106d71bb52f10164dc020f73a0c47e05171f1cb" exitCode=137 Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.133231 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c34875cbcb668d66fad6cff775a1cee539b0df1ff1d577d2d86a83805a638370" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.158616 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f7dbc7e1ee9c187a863ef9b473fad27b/startup-monitor/0.log" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.158758 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.161162 5117 status_manager.go:895] "Failed to get status for pod" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="pods \"kube-apiserver-startup-monitor-crc\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-kube-apiserver\": no relationship found between node 'crc' and this object" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.201839 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.201955 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.202195 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.202288 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.202389 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.202531 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests" (OuterVolumeSpecName: "manifests") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.202556 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.202624 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log" (OuterVolumeSpecName: "var-log") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.202684 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock" (OuterVolumeSpecName: "var-lock") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.202951 5117 reconciler_common.go:299] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.202990 5117 reconciler_common.go:299] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.203012 5117 reconciler_common.go:299] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.203031 5117 reconciler_common.go:299] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.214616 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.304806 5117 reconciler_common.go:299] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.718330 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:15:54 crc kubenswrapper[5117]: I0321 00:15:54.718427 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:15:55 crc kubenswrapper[5117]: I0321 00:15:55.144336 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 21 00:15:55 crc kubenswrapper[5117]: I0321 00:15:55.174584 5117 status_manager.go:895] "Failed to get status for pod" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="pods \"kube-apiserver-startup-monitor-crc\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-kube-apiserver\": no relationship found between node 'crc' and this object" Mar 21 00:15:55 crc kubenswrapper[5117]: I0321 00:15:55.587204 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" path="/var/lib/kubelet/pods/f7dbc7e1ee9c187a863ef9b473fad27b/volumes" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.199047 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567536-87psb"] Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.201963 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" containerName="installer" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.202010 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" containerName="installer" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.202083 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.202098 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.203095 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.203149 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="7739113f-88a1-4f1e-9fcf-d4045a611ed1" containerName="installer" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.212340 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567536-87psb"] Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.212645 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567536-87psb" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.217004 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.217983 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.219453 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.300945 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvdl5\" (UniqueName: \"kubernetes.io/projected/877a7600-33d3-446f-b83f-459c6a36cc69-kube-api-access-lvdl5\") pod \"auto-csr-approver-29567536-87psb\" (UID: \"877a7600-33d3-446f-b83f-459c6a36cc69\") " pod="openshift-infra/auto-csr-approver-29567536-87psb" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.402154 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lvdl5\" (UniqueName: \"kubernetes.io/projected/877a7600-33d3-446f-b83f-459c6a36cc69-kube-api-access-lvdl5\") pod \"auto-csr-approver-29567536-87psb\" (UID: \"877a7600-33d3-446f-b83f-459c6a36cc69\") " pod="openshift-infra/auto-csr-approver-29567536-87psb" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.426002 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvdl5\" (UniqueName: \"kubernetes.io/projected/877a7600-33d3-446f-b83f-459c6a36cc69-kube-api-access-lvdl5\") pod \"auto-csr-approver-29567536-87psb\" (UID: \"877a7600-33d3-446f-b83f-459c6a36cc69\") " pod="openshift-infra/auto-csr-approver-29567536-87psb" Mar 21 00:16:00 crc kubenswrapper[5117]: I0321 00:16:00.535457 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567536-87psb" Mar 21 00:16:01 crc kubenswrapper[5117]: I0321 00:16:01.061389 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567536-87psb"] Mar 21 00:16:01 crc kubenswrapper[5117]: I0321 00:16:01.195308 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567536-87psb" event={"ID":"877a7600-33d3-446f-b83f-459c6a36cc69","Type":"ContainerStarted","Data":"69ea5448e63b0d7a915f9190edd689eed75a20b642e9f6854a396bf037815dfd"} Mar 21 00:16:03 crc kubenswrapper[5117]: I0321 00:16:03.213825 5117 generic.go:358] "Generic (PLEG): container finished" podID="877a7600-33d3-446f-b83f-459c6a36cc69" containerID="6d07328e753842f8bfde46d56a802751ad4f1670ad572ded18d108d02ec382d5" exitCode=0 Mar 21 00:16:03 crc kubenswrapper[5117]: I0321 00:16:03.213946 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567536-87psb" event={"ID":"877a7600-33d3-446f-b83f-459c6a36cc69","Type":"ContainerDied","Data":"6d07328e753842f8bfde46d56a802751ad4f1670ad572ded18d108d02ec382d5"} Mar 21 00:16:04 crc kubenswrapper[5117]: I0321 00:16:04.522631 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567536-87psb" Mar 21 00:16:04 crc kubenswrapper[5117]: I0321 00:16:04.661006 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvdl5\" (UniqueName: \"kubernetes.io/projected/877a7600-33d3-446f-b83f-459c6a36cc69-kube-api-access-lvdl5\") pod \"877a7600-33d3-446f-b83f-459c6a36cc69\" (UID: \"877a7600-33d3-446f-b83f-459c6a36cc69\") " Mar 21 00:16:04 crc kubenswrapper[5117]: I0321 00:16:04.669435 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/877a7600-33d3-446f-b83f-459c6a36cc69-kube-api-access-lvdl5" (OuterVolumeSpecName: "kube-api-access-lvdl5") pod "877a7600-33d3-446f-b83f-459c6a36cc69" (UID: "877a7600-33d3-446f-b83f-459c6a36cc69"). InnerVolumeSpecName "kube-api-access-lvdl5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:16:04 crc kubenswrapper[5117]: I0321 00:16:04.762751 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lvdl5\" (UniqueName: \"kubernetes.io/projected/877a7600-33d3-446f-b83f-459c6a36cc69-kube-api-access-lvdl5\") on node \"crc\" DevicePath \"\"" Mar 21 00:16:05 crc kubenswrapper[5117]: I0321 00:16:05.230586 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567536-87psb" Mar 21 00:16:05 crc kubenswrapper[5117]: I0321 00:16:05.230629 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567536-87psb" event={"ID":"877a7600-33d3-446f-b83f-459c6a36cc69","Type":"ContainerDied","Data":"69ea5448e63b0d7a915f9190edd689eed75a20b642e9f6854a396bf037815dfd"} Mar 21 00:16:05 crc kubenswrapper[5117]: I0321 00:16:05.230689 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69ea5448e63b0d7a915f9190edd689eed75a20b642e9f6854a396bf037815dfd" Mar 21 00:16:19 crc kubenswrapper[5117]: I0321 00:16:19.336772 5117 generic.go:358] "Generic (PLEG): container finished" podID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerID="6b151359ff5811031ea25e18111f24eb4c41c6ba56051a143b4e52c08a57af35" exitCode=0 Mar 21 00:16:19 crc kubenswrapper[5117]: I0321 00:16:19.336882 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" event={"ID":"af2a44fd-39a9-41fc-9d1f-640e828d6a16","Type":"ContainerDied","Data":"6b151359ff5811031ea25e18111f24eb4c41c6ba56051a143b4e52c08a57af35"} Mar 21 00:16:19 crc kubenswrapper[5117]: I0321 00:16:19.338411 5117 scope.go:117] "RemoveContainer" containerID="6b151359ff5811031ea25e18111f24eb4c41c6ba56051a143b4e52c08a57af35" Mar 21 00:16:20 crc kubenswrapper[5117]: I0321 00:16:20.347982 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" event={"ID":"af2a44fd-39a9-41fc-9d1f-640e828d6a16","Type":"ContainerStarted","Data":"33ee83d92cc7b3bed271e0530269965343b85f9aa26141429c94beb16d7cb926"} Mar 21 00:16:20 crc kubenswrapper[5117]: I0321 00:16:20.348967 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:16:20 crc kubenswrapper[5117]: I0321 00:16:20.353193 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:16:24 crc kubenswrapper[5117]: I0321 00:16:24.719221 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:16:24 crc kubenswrapper[5117]: I0321 00:16:24.720168 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:16:24 crc kubenswrapper[5117]: I0321 00:16:24.720309 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:16:24 crc kubenswrapper[5117]: I0321 00:16:24.721346 5117 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4c8552401b305d82d54f228c0905aa24fe71ef36fb91e0372632f7b632cbf5e9"} pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 00:16:24 crc kubenswrapper[5117]: I0321 00:16:24.721419 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" containerID="cri-o://4c8552401b305d82d54f228c0905aa24fe71ef36fb91e0372632f7b632cbf5e9" gracePeriod=600 Mar 21 00:16:25 crc kubenswrapper[5117]: I0321 00:16:25.417973 5117 generic.go:358] "Generic (PLEG): container finished" podID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerID="4c8552401b305d82d54f228c0905aa24fe71ef36fb91e0372632f7b632cbf5e9" exitCode=0 Mar 21 00:16:25 crc kubenswrapper[5117]: I0321 00:16:25.418090 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerDied","Data":"4c8552401b305d82d54f228c0905aa24fe71ef36fb91e0372632f7b632cbf5e9"} Mar 21 00:16:25 crc kubenswrapper[5117]: I0321 00:16:25.418172 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerStarted","Data":"83c05081a1abbd89c7dc06294900b198f6c61bbe00ae5049cd1d47115a11e0ea"} Mar 21 00:16:27 crc kubenswrapper[5117]: I0321 00:16:27.733571 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f7dbc7e1ee9c187a863ef9b473fad27b/startup-monitor/0.log" Mar 21 00:16:27 crc kubenswrapper[5117]: I0321 00:16:27.753643 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f7dbc7e1ee9c187a863ef9b473fad27b/startup-monitor/0.log" Mar 21 00:16:27 crc kubenswrapper[5117]: I0321 00:16:27.824053 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:16:27 crc kubenswrapper[5117]: I0321 00:16:27.833853 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:17:01 crc kubenswrapper[5117]: I0321 00:17:01.284492 5117 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.400083 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lgn8x"] Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.401289 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lgn8x" podUID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerName="registry-server" containerID="cri-o://fb1fd2789f4cb7019f8780db05055d4abaaf9b385a76e910e3aaba35a9cd48df" gracePeriod=30 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.405533 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pqkbn"] Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.406112 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pqkbn" podUID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerName="registry-server" containerID="cri-o://250f928cbfbe573d75421660ff381149b350fb2932ca146f21b1bf2224c43eef" gracePeriod=30 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.442189 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-ll4kg"] Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.442698 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" containerID="cri-o://33ee83d92cc7b3bed271e0530269965343b85f9aa26141429c94beb16d7cb926" gracePeriod=30 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.463973 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn6n6"] Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.464579 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bn6n6" podUID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerName="registry-server" containerID="cri-o://7b584406c0cc2175e43e4ba73514501a4feb96a61c59a40c9fb6a04b8c19b657" gracePeriod=30 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.468208 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d2v8t"] Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.468700 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d2v8t" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerName="registry-server" containerID="cri-o://c60254e7c5eeb03728b5c9bf521ddf0fc6be4adb1d91a090ddec51ceac589d24" gracePeriod=30 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.470996 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-p5w88"] Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.471730 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="877a7600-33d3-446f-b83f-459c6a36cc69" containerName="oc" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.471754 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="877a7600-33d3-446f-b83f-459c6a36cc69" containerName="oc" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.471921 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="877a7600-33d3-446f-b83f-459c6a36cc69" containerName="oc" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.493623 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-p5w88"] Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.493716 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.676487 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.676603 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.676659 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-tmp\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.676733 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pvl5\" (UniqueName: \"kubernetes.io/projected/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-kube-api-access-8pvl5\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.778521 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8pvl5\" (UniqueName: \"kubernetes.io/projected/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-kube-api-access-8pvl5\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.779151 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.779218 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.779284 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-tmp\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.779948 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-tmp\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.785813 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.791509 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.806055 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pvl5\" (UniqueName: \"kubernetes.io/projected/5beaa12a-d1ce-456a-a6f4-2d9b610c540f-kube-api-access-8pvl5\") pod \"marketplace-operator-547dbd544d-p5w88\" (UID: \"5beaa12a-d1ce-456a-a6f4-2d9b610c540f\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.806532 5117 generic.go:358] "Generic (PLEG): container finished" podID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerID="33ee83d92cc7b3bed271e0530269965343b85f9aa26141429c94beb16d7cb926" exitCode=0 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.806741 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" event={"ID":"af2a44fd-39a9-41fc-9d1f-640e828d6a16","Type":"ContainerDied","Data":"33ee83d92cc7b3bed271e0530269965343b85f9aa26141429c94beb16d7cb926"} Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.806822 5117 scope.go:117] "RemoveContainer" containerID="6b151359ff5811031ea25e18111f24eb4c41c6ba56051a143b4e52c08a57af35" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.813076 5117 generic.go:358] "Generic (PLEG): container finished" podID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerID="fb1fd2789f4cb7019f8780db05055d4abaaf9b385a76e910e3aaba35a9cd48df" exitCode=0 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.813192 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lgn8x" event={"ID":"1eac1109-f55b-4a75-9d83-ad0aba830baa","Type":"ContainerDied","Data":"fb1fd2789f4cb7019f8780db05055d4abaaf9b385a76e910e3aaba35a9cd48df"} Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.813215 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lgn8x" event={"ID":"1eac1109-f55b-4a75-9d83-ad0aba830baa","Type":"ContainerDied","Data":"39b8f7e2f12cba8583afe76ba7a3d20856e9ebebeac3a689cb926030853aedf7"} Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.813225 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39b8f7e2f12cba8583afe76ba7a3d20856e9ebebeac3a689cb926030853aedf7" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.821420 5117 generic.go:358] "Generic (PLEG): container finished" podID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerID="250f928cbfbe573d75421660ff381149b350fb2932ca146f21b1bf2224c43eef" exitCode=0 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.821571 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqkbn" event={"ID":"a106308f-08b0-4f04-a431-8aa7042e4fbb","Type":"ContainerDied","Data":"250f928cbfbe573d75421660ff381149b350fb2932ca146f21b1bf2224c43eef"} Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.824390 5117 generic.go:358] "Generic (PLEG): container finished" podID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerID="7b584406c0cc2175e43e4ba73514501a4feb96a61c59a40c9fb6a04b8c19b657" exitCode=0 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.824497 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn6n6" event={"ID":"0c19bd60-8272-4e5e-81d3-10660df3c5ed","Type":"ContainerDied","Data":"7b584406c0cc2175e43e4ba73514501a4feb96a61c59a40c9fb6a04b8c19b657"} Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.832236 5117 generic.go:358] "Generic (PLEG): container finished" podID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerID="c60254e7c5eeb03728b5c9bf521ddf0fc6be4adb1d91a090ddec51ceac589d24" exitCode=0 Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.832388 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2v8t" event={"ID":"0c7aa7ad-faf4-41a3-82a1-c064c8109629","Type":"ContainerDied","Data":"c60254e7c5eeb03728b5c9bf521ddf0fc6be4adb1d91a090ddec51ceac589d24"} Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.886477 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.900597 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.905223 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.943858 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.987705 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:17:16 crc kubenswrapper[5117]: I0321 00:17:16.998068 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084051 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-catalog-content\") pod \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084147 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-utilities\") pod \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084206 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhfmx\" (UniqueName: \"kubernetes.io/projected/a106308f-08b0-4f04-a431-8aa7042e4fbb-kube-api-access-xhfmx\") pod \"a106308f-08b0-4f04-a431-8aa7042e4fbb\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084230 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-catalog-content\") pod \"a106308f-08b0-4f04-a431-8aa7042e4fbb\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084276 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgcmd\" (UniqueName: \"kubernetes.io/projected/0c7aa7ad-faf4-41a3-82a1-c064c8109629-kube-api-access-fgcmd\") pod \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084309 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfq24\" (UniqueName: \"kubernetes.io/projected/0c19bd60-8272-4e5e-81d3-10660df3c5ed-kube-api-access-dfq24\") pod \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084403 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-utilities\") pod \"a106308f-08b0-4f04-a431-8aa7042e4fbb\" (UID: \"a106308f-08b0-4f04-a431-8aa7042e4fbb\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084454 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rdvl\" (UniqueName: \"kubernetes.io/projected/1eac1109-f55b-4a75-9d83-ad0aba830baa-kube-api-access-4rdvl\") pod \"1eac1109-f55b-4a75-9d83-ad0aba830baa\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084509 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-utilities\") pod \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\" (UID: \"0c7aa7ad-faf4-41a3-82a1-c064c8109629\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084533 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-utilities\") pod \"1eac1109-f55b-4a75-9d83-ad0aba830baa\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084556 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-catalog-content\") pod \"1eac1109-f55b-4a75-9d83-ad0aba830baa\" (UID: \"1eac1109-f55b-4a75-9d83-ad0aba830baa\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.084596 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-catalog-content\") pod \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\" (UID: \"0c19bd60-8272-4e5e-81d3-10660df3c5ed\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.085651 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-utilities" (OuterVolumeSpecName: "utilities") pod "0c19bd60-8272-4e5e-81d3-10660df3c5ed" (UID: "0c19bd60-8272-4e5e-81d3-10660df3c5ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.086820 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-utilities" (OuterVolumeSpecName: "utilities") pod "a106308f-08b0-4f04-a431-8aa7042e4fbb" (UID: "a106308f-08b0-4f04-a431-8aa7042e4fbb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.087929 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-utilities" (OuterVolumeSpecName: "utilities") pod "1eac1109-f55b-4a75-9d83-ad0aba830baa" (UID: "1eac1109-f55b-4a75-9d83-ad0aba830baa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.096323 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-utilities" (OuterVolumeSpecName: "utilities") pod "0c7aa7ad-faf4-41a3-82a1-c064c8109629" (UID: "0c7aa7ad-faf4-41a3-82a1-c064c8109629"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.102587 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a106308f-08b0-4f04-a431-8aa7042e4fbb-kube-api-access-xhfmx" (OuterVolumeSpecName: "kube-api-access-xhfmx") pod "a106308f-08b0-4f04-a431-8aa7042e4fbb" (UID: "a106308f-08b0-4f04-a431-8aa7042e4fbb"). InnerVolumeSpecName "kube-api-access-xhfmx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.102620 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eac1109-f55b-4a75-9d83-ad0aba830baa-kube-api-access-4rdvl" (OuterVolumeSpecName: "kube-api-access-4rdvl") pod "1eac1109-f55b-4a75-9d83-ad0aba830baa" (UID: "1eac1109-f55b-4a75-9d83-ad0aba830baa"). InnerVolumeSpecName "kube-api-access-4rdvl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.102682 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c19bd60-8272-4e5e-81d3-10660df3c5ed-kube-api-access-dfq24" (OuterVolumeSpecName: "kube-api-access-dfq24") pod "0c19bd60-8272-4e5e-81d3-10660df3c5ed" (UID: "0c19bd60-8272-4e5e-81d3-10660df3c5ed"). InnerVolumeSpecName "kube-api-access-dfq24". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.102702 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c7aa7ad-faf4-41a3-82a1-c064c8109629-kube-api-access-fgcmd" (OuterVolumeSpecName: "kube-api-access-fgcmd") pod "0c7aa7ad-faf4-41a3-82a1-c064c8109629" (UID: "0c7aa7ad-faf4-41a3-82a1-c064c8109629"). InnerVolumeSpecName "kube-api-access-fgcmd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.127665 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c19bd60-8272-4e5e-81d3-10660df3c5ed" (UID: "0c19bd60-8272-4e5e-81d3-10660df3c5ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.139109 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1eac1109-f55b-4a75-9d83-ad0aba830baa" (UID: "1eac1109-f55b-4a75-9d83-ad0aba830baa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.179118 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a106308f-08b0-4f04-a431-8aa7042e4fbb" (UID: "a106308f-08b0-4f04-a431-8aa7042e4fbb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186324 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-trusted-ca\") pod \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186375 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/af2a44fd-39a9-41fc-9d1f-640e828d6a16-tmp\") pod \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186563 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-operator-metrics\") pod \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186589 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g8l7\" (UniqueName: \"kubernetes.io/projected/af2a44fd-39a9-41fc-9d1f-640e828d6a16-kube-api-access-9g8l7\") pod \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\" (UID: \"af2a44fd-39a9-41fc-9d1f-640e828d6a16\") " Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186747 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186765 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186774 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1eac1109-f55b-4a75-9d83-ad0aba830baa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186786 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186794 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c19bd60-8272-4e5e-81d3-10660df3c5ed-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186802 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xhfmx\" (UniqueName: \"kubernetes.io/projected/a106308f-08b0-4f04-a431-8aa7042e4fbb-kube-api-access-xhfmx\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186811 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186819 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-fgcmd\" (UniqueName: \"kubernetes.io/projected/0c7aa7ad-faf4-41a3-82a1-c064c8109629-kube-api-access-fgcmd\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186835 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dfq24\" (UniqueName: \"kubernetes.io/projected/0c19bd60-8272-4e5e-81d3-10660df3c5ed-kube-api-access-dfq24\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186844 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a106308f-08b0-4f04-a431-8aa7042e4fbb-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.186851 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4rdvl\" (UniqueName: \"kubernetes.io/projected/1eac1109-f55b-4a75-9d83-ad0aba830baa-kube-api-access-4rdvl\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.187411 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af2a44fd-39a9-41fc-9d1f-640e828d6a16-tmp" (OuterVolumeSpecName: "tmp") pod "af2a44fd-39a9-41fc-9d1f-640e828d6a16" (UID: "af2a44fd-39a9-41fc-9d1f-640e828d6a16"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.187932 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "af2a44fd-39a9-41fc-9d1f-640e828d6a16" (UID: "af2a44fd-39a9-41fc-9d1f-640e828d6a16"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.191945 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "af2a44fd-39a9-41fc-9d1f-640e828d6a16" (UID: "af2a44fd-39a9-41fc-9d1f-640e828d6a16"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.193068 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af2a44fd-39a9-41fc-9d1f-640e828d6a16-kube-api-access-9g8l7" (OuterVolumeSpecName: "kube-api-access-9g8l7") pod "af2a44fd-39a9-41fc-9d1f-640e828d6a16" (UID: "af2a44fd-39a9-41fc-9d1f-640e828d6a16"). InnerVolumeSpecName "kube-api-access-9g8l7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.233078 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c7aa7ad-faf4-41a3-82a1-c064c8109629" (UID: "0c7aa7ad-faf4-41a3-82a1-c064c8109629"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.288689 5117 reconciler_common.go:299] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.288758 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9g8l7\" (UniqueName: \"kubernetes.io/projected/af2a44fd-39a9-41fc-9d1f-640e828d6a16-kube-api-access-9g8l7\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.288782 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7aa7ad-faf4-41a3-82a1-c064c8109629-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.288800 5117 reconciler_common.go:299] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af2a44fd-39a9-41fc-9d1f-640e828d6a16-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.288819 5117 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/af2a44fd-39a9-41fc-9d1f-640e828d6a16-tmp\") on node \"crc\" DevicePath \"\"" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.375689 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-p5w88"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.395124 5117 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.842111 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqkbn" event={"ID":"a106308f-08b0-4f04-a431-8aa7042e4fbb","Type":"ContainerDied","Data":"4a9cf04658d3d2f0b8eef892a4b5461f0d31740e15bc853103f6f9bf17def73b"} Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.842135 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pqkbn" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.842197 5117 scope.go:117] "RemoveContainer" containerID="250f928cbfbe573d75421660ff381149b350fb2932ca146f21b1bf2224c43eef" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.848605 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn6n6" event={"ID":"0c19bd60-8272-4e5e-81d3-10660df3c5ed","Type":"ContainerDied","Data":"8bb2135785bcf446f3d83ca4d200e1483d2eae7d5e09afbb41c81d57f23df164"} Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.848752 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bn6n6" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.853263 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2v8t" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.853436 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2v8t" event={"ID":"0c7aa7ad-faf4-41a3-82a1-c064c8109629","Type":"ContainerDied","Data":"48bc262aea466c0bd201594f8a2d9794775a3cac5a695d0cba6dd51a62b10ec7"} Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.856586 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" event={"ID":"5beaa12a-d1ce-456a-a6f4-2d9b610c540f","Type":"ContainerStarted","Data":"dd01a29db9d23521b48490671044618c4f79321e2b3f3e41daa515a5be7879ba"} Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.856620 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" event={"ID":"5beaa12a-d1ce-456a-a6f4-2d9b610c540f","Type":"ContainerStarted","Data":"a7093ae1df79f083a5ede192c55365d48f50938af61837e299be8f67674892e7"} Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.859588 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.859815 5117 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-p5w88 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.71:8080/healthz\": dial tcp 10.217.0.71:8080: connect: connection refused" start-of-body= Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.859887 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" podUID="5beaa12a-d1ce-456a-a6f4-2d9b610c540f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.71:8080/healthz\": dial tcp 10.217.0.71:8080: connect: connection refused" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.869856 5117 scope.go:117] "RemoveContainer" containerID="b254996c08a74ae7d0a586b1ce8b34673c9b8de5d9bc0a68c55b18f1da5432e8" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.873509 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lgn8x" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.874034 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" event={"ID":"af2a44fd-39a9-41fc-9d1f-640e828d6a16","Type":"ContainerDied","Data":"0fedcbe439338cc00a1ddeb8853f641888f8c1bff6b0647aecc8b73e3e30dc05"} Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.874145 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-ll4kg" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.888340 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pqkbn"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.911558 5117 scope.go:117] "RemoveContainer" containerID="20a499d64717d049e9c7ecb04ace9ed533bc0605d630efbd78ca9661844d2ed9" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.914750 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pqkbn"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.919967 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn6n6"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.925797 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn6n6"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.940531 5117 scope.go:117] "RemoveContainer" containerID="7b584406c0cc2175e43e4ba73514501a4feb96a61c59a40c9fb6a04b8c19b657" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.953003 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d2v8t"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.964015 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" podStartSLOduration=1.9639873620000001 podStartE2EDuration="1.963987362s" podCreationTimestamp="2026-03-21 00:17:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:17:17.932139004 +0000 UTC m=+351.225426186" watchObservedRunningTime="2026-03-21 00:17:17.963987362 +0000 UTC m=+351.257274534" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.964396 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d2v8t"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.969158 5117 scope.go:117] "RemoveContainer" containerID="e74f027a89bba3cb5cc561e52ac36add0f5d4840e394010297464154ecbf26fb" Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.973494 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lgn8x"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.980011 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lgn8x"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.984084 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-ll4kg"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.988714 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-ll4kg"] Mar 21 00:17:17 crc kubenswrapper[5117]: I0321 00:17:17.990395 5117 scope.go:117] "RemoveContainer" containerID="4f1d6da7c02f7a39bc544819646368d43d4cfa5dd5aa2ccfe427a4c19d74a6aa" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.005655 5117 scope.go:117] "RemoveContainer" containerID="c60254e7c5eeb03728b5c9bf521ddf0fc6be4adb1d91a090ddec51ceac589d24" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.020908 5117 scope.go:117] "RemoveContainer" containerID="81b5faf2fb3f54852f02594fe0d6888d914a0752b0d59ab5476442da7de5e373" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.056124 5117 scope.go:117] "RemoveContainer" containerID="8fb837ac163667d358eb90e07958df816b88628d84ad80f6ac0aa2204898fdeb" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.085085 5117 scope.go:117] "RemoveContainer" containerID="33ee83d92cc7b3bed271e0530269965343b85f9aa26141429c94beb16d7cb926" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.388935 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hcbrw"] Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389572 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerName="extract-content" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389593 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerName="extract-content" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389608 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389615 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389622 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389628 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389636 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerName="extract-utilities" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389642 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerName="extract-utilities" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389650 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389656 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389669 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerName="extract-content" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389674 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerName="extract-content" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389684 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerName="extract-utilities" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389690 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerName="extract-utilities" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389700 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerName="extract-utilities" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389708 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerName="extract-utilities" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389719 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerName="extract-content" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389726 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerName="extract-content" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389740 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389748 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389756 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerName="extract-content" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389762 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerName="extract-content" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389769 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389775 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389784 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerName="extract-utilities" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389790 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerName="extract-utilities" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389801 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389807 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389909 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389920 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389929 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="1eac1109-f55b-4a75-9d83-ad0aba830baa" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389938 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" containerName="marketplace-operator" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389949 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.389957 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="a106308f-08b0-4f04-a431-8aa7042e4fbb" containerName="registry-server" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.510734 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hcbrw"] Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.510996 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.514729 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"community-operators-dockercfg-vrd5f\"" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.609814 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce1002d1-e98d-4caa-8a20-b1207c6f3ecd-catalog-content\") pod \"community-operators-hcbrw\" (UID: \"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd\") " pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.609900 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce1002d1-e98d-4caa-8a20-b1207c6f3ecd-utilities\") pod \"community-operators-hcbrw\" (UID: \"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd\") " pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.609953 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnpp8\" (UniqueName: \"kubernetes.io/projected/ce1002d1-e98d-4caa-8a20-b1207c6f3ecd-kube-api-access-lnpp8\") pod \"community-operators-hcbrw\" (UID: \"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd\") " pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.710885 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce1002d1-e98d-4caa-8a20-b1207c6f3ecd-catalog-content\") pod \"community-operators-hcbrw\" (UID: \"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd\") " pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.711048 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce1002d1-e98d-4caa-8a20-b1207c6f3ecd-utilities\") pod \"community-operators-hcbrw\" (UID: \"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd\") " pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.711150 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lnpp8\" (UniqueName: \"kubernetes.io/projected/ce1002d1-e98d-4caa-8a20-b1207c6f3ecd-kube-api-access-lnpp8\") pod \"community-operators-hcbrw\" (UID: \"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd\") " pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.711964 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce1002d1-e98d-4caa-8a20-b1207c6f3ecd-catalog-content\") pod \"community-operators-hcbrw\" (UID: \"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd\") " pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.712336 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce1002d1-e98d-4caa-8a20-b1207c6f3ecd-utilities\") pod \"community-operators-hcbrw\" (UID: \"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd\") " pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.739724 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnpp8\" (UniqueName: \"kubernetes.io/projected/ce1002d1-e98d-4caa-8a20-b1207c6f3ecd-kube-api-access-lnpp8\") pod \"community-operators-hcbrw\" (UID: \"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd\") " pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.829703 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:18 crc kubenswrapper[5117]: I0321 00:17:18.911608 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-p5w88" Mar 21 00:17:19 crc kubenswrapper[5117]: I0321 00:17:19.089264 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hcbrw"] Mar 21 00:17:19 crc kubenswrapper[5117]: I0321 00:17:19.589789 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c19bd60-8272-4e5e-81d3-10660df3c5ed" path="/var/lib/kubelet/pods/0c19bd60-8272-4e5e-81d3-10660df3c5ed/volumes" Mar 21 00:17:19 crc kubenswrapper[5117]: I0321 00:17:19.591130 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c7aa7ad-faf4-41a3-82a1-c064c8109629" path="/var/lib/kubelet/pods/0c7aa7ad-faf4-41a3-82a1-c064c8109629/volumes" Mar 21 00:17:19 crc kubenswrapper[5117]: I0321 00:17:19.591946 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eac1109-f55b-4a75-9d83-ad0aba830baa" path="/var/lib/kubelet/pods/1eac1109-f55b-4a75-9d83-ad0aba830baa/volumes" Mar 21 00:17:19 crc kubenswrapper[5117]: I0321 00:17:19.593482 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a106308f-08b0-4f04-a431-8aa7042e4fbb" path="/var/lib/kubelet/pods/a106308f-08b0-4f04-a431-8aa7042e4fbb/volumes" Mar 21 00:17:19 crc kubenswrapper[5117]: I0321 00:17:19.594516 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af2a44fd-39a9-41fc-9d1f-640e828d6a16" path="/var/lib/kubelet/pods/af2a44fd-39a9-41fc-9d1f-640e828d6a16/volumes" Mar 21 00:17:19 crc kubenswrapper[5117]: I0321 00:17:19.916284 5117 generic.go:358] "Generic (PLEG): container finished" podID="ce1002d1-e98d-4caa-8a20-b1207c6f3ecd" containerID="1b37a20442547f4abc53f64a3a257c98462e053d3886b17629151274f56e5bcf" exitCode=0 Mar 21 00:17:19 crc kubenswrapper[5117]: I0321 00:17:19.916530 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcbrw" event={"ID":"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd","Type":"ContainerDied","Data":"1b37a20442547f4abc53f64a3a257c98462e053d3886b17629151274f56e5bcf"} Mar 21 00:17:19 crc kubenswrapper[5117]: I0321 00:17:19.916619 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcbrw" event={"ID":"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd","Type":"ContainerStarted","Data":"5eb1bbdf5fabe49bb3096daa88aed8a7c7898af91eaba4a719be73731c2ed106"} Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.187844 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8xr9n"] Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.197810 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.199893 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-gg4w7\"" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.203650 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xr9n"] Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.232031 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-catalog-content\") pod \"redhat-marketplace-8xr9n\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.232150 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-utilities\") pod \"redhat-marketplace-8xr9n\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.232186 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td66s\" (UniqueName: \"kubernetes.io/projected/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-kube-api-access-td66s\") pod \"redhat-marketplace-8xr9n\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.333618 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-catalog-content\") pod \"redhat-marketplace-8xr9n\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.333718 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-utilities\") pod \"redhat-marketplace-8xr9n\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.333758 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-td66s\" (UniqueName: \"kubernetes.io/projected/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-kube-api-access-td66s\") pod \"redhat-marketplace-8xr9n\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.334180 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-utilities\") pod \"redhat-marketplace-8xr9n\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.334378 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-catalog-content\") pod \"redhat-marketplace-8xr9n\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.360148 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-td66s\" (UniqueName: \"kubernetes.io/projected/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-kube-api-access-td66s\") pod \"redhat-marketplace-8xr9n\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.523650 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.796963 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6qtzc"] Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.809500 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xr9n"] Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.809734 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.813305 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-9gxlh\"" Mar 21 00:17:20 crc kubenswrapper[5117]: W0321 00:17:20.825380 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd91c7d7a_4c8e_454c_ae77_791050b6c6e3.slice/crio-94c94d1e12072ce0b3ed8c3fbd500f7825120329d9a7648a73ff9490a0eb2ab1 WatchSource:0}: Error finding container 94c94d1e12072ce0b3ed8c3fbd500f7825120329d9a7648a73ff9490a0eb2ab1: Status 404 returned error can't find the container with id 94c94d1e12072ce0b3ed8c3fbd500f7825120329d9a7648a73ff9490a0eb2ab1 Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.839046 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6qtzc"] Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.841338 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c8a935-8d54-4f0d-acb0-8ce00c78a95e-utilities\") pod \"redhat-operators-6qtzc\" (UID: \"38c8a935-8d54-4f0d-acb0-8ce00c78a95e\") " pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.841372 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7hsn\" (UniqueName: \"kubernetes.io/projected/38c8a935-8d54-4f0d-acb0-8ce00c78a95e-kube-api-access-d7hsn\") pod \"redhat-operators-6qtzc\" (UID: \"38c8a935-8d54-4f0d-acb0-8ce00c78a95e\") " pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.841432 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c8a935-8d54-4f0d-acb0-8ce00c78a95e-catalog-content\") pod \"redhat-operators-6qtzc\" (UID: \"38c8a935-8d54-4f0d-acb0-8ce00c78a95e\") " pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.924641 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcbrw" event={"ID":"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd","Type":"ContainerStarted","Data":"ab59d13f643fa3282576fa1bcc2f8e24d657e193f3f0d9cf7b75824ba2e170d6"} Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.927429 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xr9n" event={"ID":"d91c7d7a-4c8e-454c-ae77-791050b6c6e3","Type":"ContainerStarted","Data":"94c94d1e12072ce0b3ed8c3fbd500f7825120329d9a7648a73ff9490a0eb2ab1"} Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.950785 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c8a935-8d54-4f0d-acb0-8ce00c78a95e-catalog-content\") pod \"redhat-operators-6qtzc\" (UID: \"38c8a935-8d54-4f0d-acb0-8ce00c78a95e\") " pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.951136 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c8a935-8d54-4f0d-acb0-8ce00c78a95e-utilities\") pod \"redhat-operators-6qtzc\" (UID: \"38c8a935-8d54-4f0d-acb0-8ce00c78a95e\") " pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.951201 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d7hsn\" (UniqueName: \"kubernetes.io/projected/38c8a935-8d54-4f0d-acb0-8ce00c78a95e-kube-api-access-d7hsn\") pod \"redhat-operators-6qtzc\" (UID: \"38c8a935-8d54-4f0d-acb0-8ce00c78a95e\") " pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.951418 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c8a935-8d54-4f0d-acb0-8ce00c78a95e-catalog-content\") pod \"redhat-operators-6qtzc\" (UID: \"38c8a935-8d54-4f0d-acb0-8ce00c78a95e\") " pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.951550 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c8a935-8d54-4f0d-acb0-8ce00c78a95e-utilities\") pod \"redhat-operators-6qtzc\" (UID: \"38c8a935-8d54-4f0d-acb0-8ce00c78a95e\") " pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:20 crc kubenswrapper[5117]: I0321 00:17:20.975632 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7hsn\" (UniqueName: \"kubernetes.io/projected/38c8a935-8d54-4f0d-acb0-8ce00c78a95e-kube-api-access-d7hsn\") pod \"redhat-operators-6qtzc\" (UID: \"38c8a935-8d54-4f0d-acb0-8ce00c78a95e\") " pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:21 crc kubenswrapper[5117]: I0321 00:17:21.155077 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:21 crc kubenswrapper[5117]: I0321 00:17:21.391450 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6qtzc"] Mar 21 00:17:21 crc kubenswrapper[5117]: I0321 00:17:21.937388 5117 generic.go:358] "Generic (PLEG): container finished" podID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerID="4a59ead9340cf7d87dc5e3a23cda40a56f9ac880489d25f65efb954c6c46f094" exitCode=0 Mar 21 00:17:21 crc kubenswrapper[5117]: I0321 00:17:21.937561 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xr9n" event={"ID":"d91c7d7a-4c8e-454c-ae77-791050b6c6e3","Type":"ContainerDied","Data":"4a59ead9340cf7d87dc5e3a23cda40a56f9ac880489d25f65efb954c6c46f094"} Mar 21 00:17:21 crc kubenswrapper[5117]: I0321 00:17:21.941468 5117 generic.go:358] "Generic (PLEG): container finished" podID="38c8a935-8d54-4f0d-acb0-8ce00c78a95e" containerID="4f1ae7f8bceb5178f3adf116a77303ecfa5931bde82353f646c3e1f5033fb854" exitCode=0 Mar 21 00:17:21 crc kubenswrapper[5117]: I0321 00:17:21.941748 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qtzc" event={"ID":"38c8a935-8d54-4f0d-acb0-8ce00c78a95e","Type":"ContainerDied","Data":"4f1ae7f8bceb5178f3adf116a77303ecfa5931bde82353f646c3e1f5033fb854"} Mar 21 00:17:21 crc kubenswrapper[5117]: I0321 00:17:21.941787 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qtzc" event={"ID":"38c8a935-8d54-4f0d-acb0-8ce00c78a95e","Type":"ContainerStarted","Data":"e52cb4324c1f6f94c8f14f3d7726abb87d81f571f12e7864dc148f39c9e0df93"} Mar 21 00:17:21 crc kubenswrapper[5117]: I0321 00:17:21.945330 5117 generic.go:358] "Generic (PLEG): container finished" podID="ce1002d1-e98d-4caa-8a20-b1207c6f3ecd" containerID="ab59d13f643fa3282576fa1bcc2f8e24d657e193f3f0d9cf7b75824ba2e170d6" exitCode=0 Mar 21 00:17:21 crc kubenswrapper[5117]: I0321 00:17:21.945426 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcbrw" event={"ID":"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd","Type":"ContainerDied","Data":"ab59d13f643fa3282576fa1bcc2f8e24d657e193f3f0d9cf7b75824ba2e170d6"} Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.590275 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v428c"] Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.604628 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.606465 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v428c"] Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.608066 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-7cl8d\"" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.678716 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/879927af-6b88-4da1-915a-91728a3f2bd1-utilities\") pod \"certified-operators-v428c\" (UID: \"879927af-6b88-4da1-915a-91728a3f2bd1\") " pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.678781 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89rvp\" (UniqueName: \"kubernetes.io/projected/879927af-6b88-4da1-915a-91728a3f2bd1-kube-api-access-89rvp\") pod \"certified-operators-v428c\" (UID: \"879927af-6b88-4da1-915a-91728a3f2bd1\") " pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.678848 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/879927af-6b88-4da1-915a-91728a3f2bd1-catalog-content\") pod \"certified-operators-v428c\" (UID: \"879927af-6b88-4da1-915a-91728a3f2bd1\") " pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.779949 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/879927af-6b88-4da1-915a-91728a3f2bd1-utilities\") pod \"certified-operators-v428c\" (UID: \"879927af-6b88-4da1-915a-91728a3f2bd1\") " pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.780026 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-89rvp\" (UniqueName: \"kubernetes.io/projected/879927af-6b88-4da1-915a-91728a3f2bd1-kube-api-access-89rvp\") pod \"certified-operators-v428c\" (UID: \"879927af-6b88-4da1-915a-91728a3f2bd1\") " pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.780071 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/879927af-6b88-4da1-915a-91728a3f2bd1-catalog-content\") pod \"certified-operators-v428c\" (UID: \"879927af-6b88-4da1-915a-91728a3f2bd1\") " pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.780635 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/879927af-6b88-4da1-915a-91728a3f2bd1-utilities\") pod \"certified-operators-v428c\" (UID: \"879927af-6b88-4da1-915a-91728a3f2bd1\") " pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.780667 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/879927af-6b88-4da1-915a-91728a3f2bd1-catalog-content\") pod \"certified-operators-v428c\" (UID: \"879927af-6b88-4da1-915a-91728a3f2bd1\") " pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.816912 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-89rvp\" (UniqueName: \"kubernetes.io/projected/879927af-6b88-4da1-915a-91728a3f2bd1-kube-api-access-89rvp\") pod \"certified-operators-v428c\" (UID: \"879927af-6b88-4da1-915a-91728a3f2bd1\") " pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.921025 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.974474 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcbrw" event={"ID":"ce1002d1-e98d-4caa-8a20-b1207c6f3ecd","Type":"ContainerStarted","Data":"3334dec993c2631138c9d69c6e8f6daa975a86ecae8b1ce5ea5243d658209f50"} Mar 21 00:17:22 crc kubenswrapper[5117]: I0321 00:17:22.997725 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hcbrw" podStartSLOduration=4.246262995 podStartE2EDuration="4.997707139s" podCreationTimestamp="2026-03-21 00:17:18 +0000 UTC" firstStartedPulling="2026-03-21 00:17:19.91814563 +0000 UTC m=+353.211432832" lastFinishedPulling="2026-03-21 00:17:20.669589804 +0000 UTC m=+353.962876976" observedRunningTime="2026-03-21 00:17:22.99576689 +0000 UTC m=+356.289054062" watchObservedRunningTime="2026-03-21 00:17:22.997707139 +0000 UTC m=+356.290994311" Mar 21 00:17:23 crc kubenswrapper[5117]: I0321 00:17:23.453554 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v428c"] Mar 21 00:17:23 crc kubenswrapper[5117]: W0321 00:17:23.458875 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod879927af_6b88_4da1_915a_91728a3f2bd1.slice/crio-54219e7846303cfd6e9405a34afdfb83afc3c9f2ddb84bbb698321d6ea413b72 WatchSource:0}: Error finding container 54219e7846303cfd6e9405a34afdfb83afc3c9f2ddb84bbb698321d6ea413b72: Status 404 returned error can't find the container with id 54219e7846303cfd6e9405a34afdfb83afc3c9f2ddb84bbb698321d6ea413b72 Mar 21 00:17:23 crc kubenswrapper[5117]: I0321 00:17:23.981901 5117 generic.go:358] "Generic (PLEG): container finished" podID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerID="196017919a8d16cbf50f101f9284c81f847e451357e39bcbc68e85c3bae37bdf" exitCode=0 Mar 21 00:17:23 crc kubenswrapper[5117]: I0321 00:17:23.981976 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xr9n" event={"ID":"d91c7d7a-4c8e-454c-ae77-791050b6c6e3","Type":"ContainerDied","Data":"196017919a8d16cbf50f101f9284c81f847e451357e39bcbc68e85c3bae37bdf"} Mar 21 00:17:23 crc kubenswrapper[5117]: I0321 00:17:23.983962 5117 generic.go:358] "Generic (PLEG): container finished" podID="38c8a935-8d54-4f0d-acb0-8ce00c78a95e" containerID="8376c6435bf19792bd7b2d7077416b0353762e2a4bd40709c59ea4c1c1f5d2ed" exitCode=0 Mar 21 00:17:23 crc kubenswrapper[5117]: I0321 00:17:23.984043 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qtzc" event={"ID":"38c8a935-8d54-4f0d-acb0-8ce00c78a95e","Type":"ContainerDied","Data":"8376c6435bf19792bd7b2d7077416b0353762e2a4bd40709c59ea4c1c1f5d2ed"} Mar 21 00:17:23 crc kubenswrapper[5117]: I0321 00:17:23.986193 5117 generic.go:358] "Generic (PLEG): container finished" podID="879927af-6b88-4da1-915a-91728a3f2bd1" containerID="9ee870ccf30235f3042c133ccdc369e9e46f841450871997e77fddb2b5ac7854" exitCode=0 Mar 21 00:17:23 crc kubenswrapper[5117]: I0321 00:17:23.986341 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v428c" event={"ID":"879927af-6b88-4da1-915a-91728a3f2bd1","Type":"ContainerDied","Data":"9ee870ccf30235f3042c133ccdc369e9e46f841450871997e77fddb2b5ac7854"} Mar 21 00:17:23 crc kubenswrapper[5117]: I0321 00:17:23.986377 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v428c" event={"ID":"879927af-6b88-4da1-915a-91728a3f2bd1","Type":"ContainerStarted","Data":"54219e7846303cfd6e9405a34afdfb83afc3c9f2ddb84bbb698321d6ea413b72"} Mar 21 00:17:24 crc kubenswrapper[5117]: I0321 00:17:24.995368 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xr9n" event={"ID":"d91c7d7a-4c8e-454c-ae77-791050b6c6e3","Type":"ContainerStarted","Data":"baac535c8d9804bf9b83f22e50e9550ac413e25ff7243825ed49e4773dc6f84a"} Mar 21 00:17:24 crc kubenswrapper[5117]: I0321 00:17:24.998367 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qtzc" event={"ID":"38c8a935-8d54-4f0d-acb0-8ce00c78a95e","Type":"ContainerStarted","Data":"bb1bca36b2b0d2219b6cb7b95b549606fe620621776e38e72fc2335a7aa0aee0"} Mar 21 00:17:25 crc kubenswrapper[5117]: I0321 00:17:25.001861 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v428c" event={"ID":"879927af-6b88-4da1-915a-91728a3f2bd1","Type":"ContainerStarted","Data":"f7ad30811cdb3f92f4e23be00138ee5f63b8801f60e833e9b44aa90da34bcba6"} Mar 21 00:17:25 crc kubenswrapper[5117]: I0321 00:17:25.022187 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8xr9n" podStartSLOduration=4.115257268 podStartE2EDuration="5.022159737s" podCreationTimestamp="2026-03-21 00:17:20 +0000 UTC" firstStartedPulling="2026-03-21 00:17:21.938834773 +0000 UTC m=+355.232121955" lastFinishedPulling="2026-03-21 00:17:22.845737252 +0000 UTC m=+356.139024424" observedRunningTime="2026-03-21 00:17:25.02123923 +0000 UTC m=+358.314526422" watchObservedRunningTime="2026-03-21 00:17:25.022159737 +0000 UTC m=+358.315446909" Mar 21 00:17:25 crc kubenswrapper[5117]: I0321 00:17:25.069272 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6qtzc" podStartSLOduration=4.172557139 podStartE2EDuration="5.069238673s" podCreationTimestamp="2026-03-21 00:17:20 +0000 UTC" firstStartedPulling="2026-03-21 00:17:21.9426446 +0000 UTC m=+355.235931772" lastFinishedPulling="2026-03-21 00:17:22.839326134 +0000 UTC m=+356.132613306" observedRunningTime="2026-03-21 00:17:25.048877359 +0000 UTC m=+358.342164531" watchObservedRunningTime="2026-03-21 00:17:25.069238673 +0000 UTC m=+358.362525845" Mar 21 00:17:26 crc kubenswrapper[5117]: I0321 00:17:26.010695 5117 generic.go:358] "Generic (PLEG): container finished" podID="879927af-6b88-4da1-915a-91728a3f2bd1" containerID="f7ad30811cdb3f92f4e23be00138ee5f63b8801f60e833e9b44aa90da34bcba6" exitCode=0 Mar 21 00:17:26 crc kubenswrapper[5117]: I0321 00:17:26.010766 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v428c" event={"ID":"879927af-6b88-4da1-915a-91728a3f2bd1","Type":"ContainerDied","Data":"f7ad30811cdb3f92f4e23be00138ee5f63b8801f60e833e9b44aa90da34bcba6"} Mar 21 00:17:27 crc kubenswrapper[5117]: I0321 00:17:27.022400 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v428c" event={"ID":"879927af-6b88-4da1-915a-91728a3f2bd1","Type":"ContainerStarted","Data":"eb28149420fe8ad650080b357b7bb6ba6a576c0723aabe978a01a79791e06061"} Mar 21 00:17:27 crc kubenswrapper[5117]: I0321 00:17:27.042935 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v428c" podStartSLOduration=4.291183689 podStartE2EDuration="5.042910701s" podCreationTimestamp="2026-03-21 00:17:22 +0000 UTC" firstStartedPulling="2026-03-21 00:17:23.98703949 +0000 UTC m=+357.280326662" lastFinishedPulling="2026-03-21 00:17:24.738766482 +0000 UTC m=+358.032053674" observedRunningTime="2026-03-21 00:17:27.041921382 +0000 UTC m=+360.335208544" watchObservedRunningTime="2026-03-21 00:17:27.042910701 +0000 UTC m=+360.336197873" Mar 21 00:17:28 crc kubenswrapper[5117]: I0321 00:17:28.831052 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:28 crc kubenswrapper[5117]: I0321 00:17:28.831578 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:28 crc kubenswrapper[5117]: I0321 00:17:28.896802 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:29 crc kubenswrapper[5117]: I0321 00:17:29.073450 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hcbrw" Mar 21 00:17:30 crc kubenswrapper[5117]: I0321 00:17:30.523869 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:30 crc kubenswrapper[5117]: I0321 00:17:30.523944 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:30 crc kubenswrapper[5117]: I0321 00:17:30.563161 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:31 crc kubenswrapper[5117]: I0321 00:17:31.156061 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:31 crc kubenswrapper[5117]: I0321 00:17:31.156481 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:31 crc kubenswrapper[5117]: I0321 00:17:31.166995 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:17:31 crc kubenswrapper[5117]: I0321 00:17:31.203766 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:32 crc kubenswrapper[5117]: I0321 00:17:32.126313 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6qtzc" Mar 21 00:17:32 crc kubenswrapper[5117]: I0321 00:17:32.922006 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:32 crc kubenswrapper[5117]: I0321 00:17:32.923152 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:32 crc kubenswrapper[5117]: I0321 00:17:32.980567 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:17:33 crc kubenswrapper[5117]: I0321 00:17:33.141408 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v428c" Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.137297 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567538-2vtzm"] Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.236137 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567538-2vtzm"] Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.236346 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567538-2vtzm" Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.239182 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.239861 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.241278 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.311226 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w9zk\" (UniqueName: \"kubernetes.io/projected/18e279db-6c97-4c52-b2b6-8b8f1ad0937f-kube-api-access-6w9zk\") pod \"auto-csr-approver-29567538-2vtzm\" (UID: \"18e279db-6c97-4c52-b2b6-8b8f1ad0937f\") " pod="openshift-infra/auto-csr-approver-29567538-2vtzm" Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.412748 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6w9zk\" (UniqueName: \"kubernetes.io/projected/18e279db-6c97-4c52-b2b6-8b8f1ad0937f-kube-api-access-6w9zk\") pod \"auto-csr-approver-29567538-2vtzm\" (UID: \"18e279db-6c97-4c52-b2b6-8b8f1ad0937f\") " pod="openshift-infra/auto-csr-approver-29567538-2vtzm" Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.438785 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w9zk\" (UniqueName: \"kubernetes.io/projected/18e279db-6c97-4c52-b2b6-8b8f1ad0937f-kube-api-access-6w9zk\") pod \"auto-csr-approver-29567538-2vtzm\" (UID: \"18e279db-6c97-4c52-b2b6-8b8f1ad0937f\") " pod="openshift-infra/auto-csr-approver-29567538-2vtzm" Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.558412 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567538-2vtzm" Mar 21 00:18:00 crc kubenswrapper[5117]: I0321 00:18:00.788224 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567538-2vtzm"] Mar 21 00:18:01 crc kubenswrapper[5117]: I0321 00:18:01.290288 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567538-2vtzm" event={"ID":"18e279db-6c97-4c52-b2b6-8b8f1ad0937f","Type":"ContainerStarted","Data":"af0931449572d8dbc85d25ac1c9cbf452716d2fd3276d63a11ec924445737609"} Mar 21 00:18:02 crc kubenswrapper[5117]: I0321 00:18:02.297230 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567538-2vtzm" event={"ID":"18e279db-6c97-4c52-b2b6-8b8f1ad0937f","Type":"ContainerStarted","Data":"798478e891c0f903b943431f5d3dc0fc1a6e3e8f33b76532864796896db5578e"} Mar 21 00:18:02 crc kubenswrapper[5117]: I0321 00:18:02.329635 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567538-2vtzm" podStartSLOduration=1.426687384 podStartE2EDuration="2.329608851s" podCreationTimestamp="2026-03-21 00:18:00 +0000 UTC" firstStartedPulling="2026-03-21 00:18:00.802150398 +0000 UTC m=+394.095437570" lastFinishedPulling="2026-03-21 00:18:01.705071825 +0000 UTC m=+394.998359037" observedRunningTime="2026-03-21 00:18:02.316440086 +0000 UTC m=+395.609727258" watchObservedRunningTime="2026-03-21 00:18:02.329608851 +0000 UTC m=+395.622896033" Mar 21 00:18:03 crc kubenswrapper[5117]: I0321 00:18:03.305663 5117 generic.go:358] "Generic (PLEG): container finished" podID="18e279db-6c97-4c52-b2b6-8b8f1ad0937f" containerID="798478e891c0f903b943431f5d3dc0fc1a6e3e8f33b76532864796896db5578e" exitCode=0 Mar 21 00:18:03 crc kubenswrapper[5117]: I0321 00:18:03.305771 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567538-2vtzm" event={"ID":"18e279db-6c97-4c52-b2b6-8b8f1ad0937f","Type":"ContainerDied","Data":"798478e891c0f903b943431f5d3dc0fc1a6e3e8f33b76532864796896db5578e"} Mar 21 00:18:04 crc kubenswrapper[5117]: I0321 00:18:04.635878 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567538-2vtzm" Mar 21 00:18:04 crc kubenswrapper[5117]: I0321 00:18:04.684769 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w9zk\" (UniqueName: \"kubernetes.io/projected/18e279db-6c97-4c52-b2b6-8b8f1ad0937f-kube-api-access-6w9zk\") pod \"18e279db-6c97-4c52-b2b6-8b8f1ad0937f\" (UID: \"18e279db-6c97-4c52-b2b6-8b8f1ad0937f\") " Mar 21 00:18:04 crc kubenswrapper[5117]: I0321 00:18:04.692341 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18e279db-6c97-4c52-b2b6-8b8f1ad0937f-kube-api-access-6w9zk" (OuterVolumeSpecName: "kube-api-access-6w9zk") pod "18e279db-6c97-4c52-b2b6-8b8f1ad0937f" (UID: "18e279db-6c97-4c52-b2b6-8b8f1ad0937f"). InnerVolumeSpecName "kube-api-access-6w9zk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:18:04 crc kubenswrapper[5117]: I0321 00:18:04.786811 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6w9zk\" (UniqueName: \"kubernetes.io/projected/18e279db-6c97-4c52-b2b6-8b8f1ad0937f-kube-api-access-6w9zk\") on node \"crc\" DevicePath \"\"" Mar 21 00:18:05 crc kubenswrapper[5117]: I0321 00:18:05.338755 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567538-2vtzm" event={"ID":"18e279db-6c97-4c52-b2b6-8b8f1ad0937f","Type":"ContainerDied","Data":"af0931449572d8dbc85d25ac1c9cbf452716d2fd3276d63a11ec924445737609"} Mar 21 00:18:05 crc kubenswrapper[5117]: I0321 00:18:05.338815 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af0931449572d8dbc85d25ac1c9cbf452716d2fd3276d63a11ec924445737609" Mar 21 00:18:05 crc kubenswrapper[5117]: I0321 00:18:05.338864 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567538-2vtzm" Mar 21 00:18:54 crc kubenswrapper[5117]: I0321 00:18:54.718157 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:18:54 crc kubenswrapper[5117]: I0321 00:18:54.719028 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:19:24 crc kubenswrapper[5117]: I0321 00:19:24.718424 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:19:24 crc kubenswrapper[5117]: I0321 00:19:24.719428 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:19:27 crc kubenswrapper[5117]: I0321 00:19:27.870904 5117 scope.go:117] "RemoveContainer" containerID="5b22728c35965a85542c89a237d68e3ff3492ac6f1a71e55c83a7d7eca7a408b" Mar 21 00:19:54 crc kubenswrapper[5117]: I0321 00:19:54.718369 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:19:54 crc kubenswrapper[5117]: I0321 00:19:54.719375 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:19:54 crc kubenswrapper[5117]: I0321 00:19:54.719458 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:19:54 crc kubenswrapper[5117]: I0321 00:19:54.720534 5117 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"83c05081a1abbd89c7dc06294900b198f6c61bbe00ae5049cd1d47115a11e0ea"} pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 00:19:54 crc kubenswrapper[5117]: I0321 00:19:54.720666 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" containerID="cri-o://83c05081a1abbd89c7dc06294900b198f6c61bbe00ae5049cd1d47115a11e0ea" gracePeriod=600 Mar 21 00:19:55 crc kubenswrapper[5117]: I0321 00:19:55.231824 5117 generic.go:358] "Generic (PLEG): container finished" podID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerID="83c05081a1abbd89c7dc06294900b198f6c61bbe00ae5049cd1d47115a11e0ea" exitCode=0 Mar 21 00:19:55 crc kubenswrapper[5117]: I0321 00:19:55.231928 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerDied","Data":"83c05081a1abbd89c7dc06294900b198f6c61bbe00ae5049cd1d47115a11e0ea"} Mar 21 00:19:55 crc kubenswrapper[5117]: I0321 00:19:55.232416 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerStarted","Data":"f2fd8729d8670a753e93005428f17279031e73d4ffabcffd4b6e8f29d6bcce8b"} Mar 21 00:19:55 crc kubenswrapper[5117]: I0321 00:19:55.232445 5117 scope.go:117] "RemoveContainer" containerID="4c8552401b305d82d54f228c0905aa24fe71ef36fb91e0372632f7b632cbf5e9" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.152805 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567540-4pg74"] Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.154654 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="18e279db-6c97-4c52-b2b6-8b8f1ad0937f" containerName="oc" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.154679 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="18e279db-6c97-4c52-b2b6-8b8f1ad0937f" containerName="oc" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.154832 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="18e279db-6c97-4c52-b2b6-8b8f1ad0937f" containerName="oc" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.165463 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567540-4pg74"] Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.165647 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567540-4pg74" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.168719 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.169108 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.169577 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.240054 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdw7j\" (UniqueName: \"kubernetes.io/projected/7c9e3f55-a1b5-45fd-989c-b9c919502c57-kube-api-access-hdw7j\") pod \"auto-csr-approver-29567540-4pg74\" (UID: \"7c9e3f55-a1b5-45fd-989c-b9c919502c57\") " pod="openshift-infra/auto-csr-approver-29567540-4pg74" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.342146 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hdw7j\" (UniqueName: \"kubernetes.io/projected/7c9e3f55-a1b5-45fd-989c-b9c919502c57-kube-api-access-hdw7j\") pod \"auto-csr-approver-29567540-4pg74\" (UID: \"7c9e3f55-a1b5-45fd-989c-b9c919502c57\") " pod="openshift-infra/auto-csr-approver-29567540-4pg74" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.372652 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdw7j\" (UniqueName: \"kubernetes.io/projected/7c9e3f55-a1b5-45fd-989c-b9c919502c57-kube-api-access-hdw7j\") pod \"auto-csr-approver-29567540-4pg74\" (UID: \"7c9e3f55-a1b5-45fd-989c-b9c919502c57\") " pod="openshift-infra/auto-csr-approver-29567540-4pg74" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.489407 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567540-4pg74" Mar 21 00:20:00 crc kubenswrapper[5117]: I0321 00:20:00.699879 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567540-4pg74"] Mar 21 00:20:01 crc kubenswrapper[5117]: I0321 00:20:01.281532 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567540-4pg74" event={"ID":"7c9e3f55-a1b5-45fd-989c-b9c919502c57","Type":"ContainerStarted","Data":"0ebaf8c3891a95cd35afe5313042b835d60fd982f6d8559b2464bf74b7e78f02"} Mar 21 00:20:03 crc kubenswrapper[5117]: I0321 00:20:03.296839 5117 generic.go:358] "Generic (PLEG): container finished" podID="7c9e3f55-a1b5-45fd-989c-b9c919502c57" containerID="e5763f91a5427b8cdd0cdc1e2e284f9c3cb0ca638a2ab7cc5c6c092f8749de22" exitCode=0 Mar 21 00:20:03 crc kubenswrapper[5117]: I0321 00:20:03.296912 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567540-4pg74" event={"ID":"7c9e3f55-a1b5-45fd-989c-b9c919502c57","Type":"ContainerDied","Data":"e5763f91a5427b8cdd0cdc1e2e284f9c3cb0ca638a2ab7cc5c6c092f8749de22"} Mar 21 00:20:04 crc kubenswrapper[5117]: I0321 00:20:04.603496 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567540-4pg74" Mar 21 00:20:04 crc kubenswrapper[5117]: I0321 00:20:04.710305 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdw7j\" (UniqueName: \"kubernetes.io/projected/7c9e3f55-a1b5-45fd-989c-b9c919502c57-kube-api-access-hdw7j\") pod \"7c9e3f55-a1b5-45fd-989c-b9c919502c57\" (UID: \"7c9e3f55-a1b5-45fd-989c-b9c919502c57\") " Mar 21 00:20:04 crc kubenswrapper[5117]: I0321 00:20:04.720435 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c9e3f55-a1b5-45fd-989c-b9c919502c57-kube-api-access-hdw7j" (OuterVolumeSpecName: "kube-api-access-hdw7j") pod "7c9e3f55-a1b5-45fd-989c-b9c919502c57" (UID: "7c9e3f55-a1b5-45fd-989c-b9c919502c57"). InnerVolumeSpecName "kube-api-access-hdw7j". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:20:04 crc kubenswrapper[5117]: I0321 00:20:04.811919 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hdw7j\" (UniqueName: \"kubernetes.io/projected/7c9e3f55-a1b5-45fd-989c-b9c919502c57-kube-api-access-hdw7j\") on node \"crc\" DevicePath \"\"" Mar 21 00:20:05 crc kubenswrapper[5117]: I0321 00:20:05.310929 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567540-4pg74" Mar 21 00:20:05 crc kubenswrapper[5117]: I0321 00:20:05.310962 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567540-4pg74" event={"ID":"7c9e3f55-a1b5-45fd-989c-b9c919502c57","Type":"ContainerDied","Data":"0ebaf8c3891a95cd35afe5313042b835d60fd982f6d8559b2464bf74b7e78f02"} Mar 21 00:20:05 crc kubenswrapper[5117]: I0321 00:20:05.311004 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ebaf8c3891a95cd35afe5313042b835d60fd982f6d8559b2464bf74b7e78f02" Mar 21 00:20:05 crc kubenswrapper[5117]: I0321 00:20:05.691811 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567534-vhz59"] Mar 21 00:20:05 crc kubenswrapper[5117]: I0321 00:20:05.698915 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567534-vhz59"] Mar 21 00:20:07 crc kubenswrapper[5117]: I0321 00:20:07.587582 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba892113-95e3-461f-a9c3-34295e210034" path="/var/lib/kubelet/pods/ba892113-95e3-461f-a9c3-34295e210034/volumes" Mar 21 00:20:27 crc kubenswrapper[5117]: I0321 00:20:27.911105 5117 scope.go:117] "RemoveContainer" containerID="5789968292dbaba1a6ec24956d0efe67faa98c185c816821c00fef9e8009bec5" Mar 21 00:20:28 crc kubenswrapper[5117]: I0321 00:20:28.020131 5117 scope.go:117] "RemoveContainer" containerID="caa984f8fd4f103859bf67dd88c6c152ef152721e4f0a0ba01b514c9adda7978" Mar 21 00:20:28 crc kubenswrapper[5117]: I0321 00:20:28.053485 5117 scope.go:117] "RemoveContainer" containerID="fb1fd2789f4cb7019f8780db05055d4abaaf9b385a76e910e3aaba35a9cd48df" Mar 21 00:21:27 crc kubenswrapper[5117]: I0321 00:21:27.940798 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:21:27 crc kubenswrapper[5117]: I0321 00:21:27.948229 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:21:28 crc kubenswrapper[5117]: I0321 00:21:28.123646 5117 scope.go:117] "RemoveContainer" containerID="5ded4898674941b2ee00c3899106d71bb52f10164dc020f73a0c47e05171f1cb" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.156656 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567542-2fs89"] Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.159464 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7c9e3f55-a1b5-45fd-989c-b9c919502c57" containerName="oc" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.159496 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c9e3f55-a1b5-45fd-989c-b9c919502c57" containerName="oc" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.159746 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="7c9e3f55-a1b5-45fd-989c-b9c919502c57" containerName="oc" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.164784 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567542-2fs89" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.167538 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.167620 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567542-2fs89"] Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.168847 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.169008 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.264961 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q764r\" (UniqueName: \"kubernetes.io/projected/ff1f8f36-82ad-4771-b8de-b5dacf30a13a-kube-api-access-q764r\") pod \"auto-csr-approver-29567542-2fs89\" (UID: \"ff1f8f36-82ad-4771-b8de-b5dacf30a13a\") " pod="openshift-infra/auto-csr-approver-29567542-2fs89" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.366617 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-q764r\" (UniqueName: \"kubernetes.io/projected/ff1f8f36-82ad-4771-b8de-b5dacf30a13a-kube-api-access-q764r\") pod \"auto-csr-approver-29567542-2fs89\" (UID: \"ff1f8f36-82ad-4771-b8de-b5dacf30a13a\") " pod="openshift-infra/auto-csr-approver-29567542-2fs89" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.401315 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-q764r\" (UniqueName: \"kubernetes.io/projected/ff1f8f36-82ad-4771-b8de-b5dacf30a13a-kube-api-access-q764r\") pod \"auto-csr-approver-29567542-2fs89\" (UID: \"ff1f8f36-82ad-4771-b8de-b5dacf30a13a\") " pod="openshift-infra/auto-csr-approver-29567542-2fs89" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.491764 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567542-2fs89" Mar 21 00:22:00 crc kubenswrapper[5117]: I0321 00:22:00.802998 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567542-2fs89"] Mar 21 00:22:01 crc kubenswrapper[5117]: I0321 00:22:01.286884 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567542-2fs89" event={"ID":"ff1f8f36-82ad-4771-b8de-b5dacf30a13a","Type":"ContainerStarted","Data":"30fb760ff15f2ac4088149fee1df147c7c1a0725bf53e57ad293eb9250b3191f"} Mar 21 00:22:02 crc kubenswrapper[5117]: I0321 00:22:02.295416 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567542-2fs89" event={"ID":"ff1f8f36-82ad-4771-b8de-b5dacf30a13a","Type":"ContainerStarted","Data":"42dfd97a193e2da73a25b9bd568728ae5b1e21c466dee8e8bb596985024ae4ec"} Mar 21 00:22:02 crc kubenswrapper[5117]: I0321 00:22:02.313519 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567542-2fs89" podStartSLOduration=1.232902601 podStartE2EDuration="2.313494799s" podCreationTimestamp="2026-03-21 00:22:00 +0000 UTC" firstStartedPulling="2026-03-21 00:22:00.813356628 +0000 UTC m=+634.106643820" lastFinishedPulling="2026-03-21 00:22:01.893948806 +0000 UTC m=+635.187236018" observedRunningTime="2026-03-21 00:22:02.31058154 +0000 UTC m=+635.603868732" watchObservedRunningTime="2026-03-21 00:22:02.313494799 +0000 UTC m=+635.606781981" Mar 21 00:22:03 crc kubenswrapper[5117]: I0321 00:22:03.307381 5117 generic.go:358] "Generic (PLEG): container finished" podID="ff1f8f36-82ad-4771-b8de-b5dacf30a13a" containerID="42dfd97a193e2da73a25b9bd568728ae5b1e21c466dee8e8bb596985024ae4ec" exitCode=0 Mar 21 00:22:03 crc kubenswrapper[5117]: I0321 00:22:03.307534 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567542-2fs89" event={"ID":"ff1f8f36-82ad-4771-b8de-b5dacf30a13a","Type":"ContainerDied","Data":"42dfd97a193e2da73a25b9bd568728ae5b1e21c466dee8e8bb596985024ae4ec"} Mar 21 00:22:04 crc kubenswrapper[5117]: I0321 00:22:04.565857 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567542-2fs89" Mar 21 00:22:04 crc kubenswrapper[5117]: I0321 00:22:04.639648 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q764r\" (UniqueName: \"kubernetes.io/projected/ff1f8f36-82ad-4771-b8de-b5dacf30a13a-kube-api-access-q764r\") pod \"ff1f8f36-82ad-4771-b8de-b5dacf30a13a\" (UID: \"ff1f8f36-82ad-4771-b8de-b5dacf30a13a\") " Mar 21 00:22:04 crc kubenswrapper[5117]: I0321 00:22:04.647239 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff1f8f36-82ad-4771-b8de-b5dacf30a13a-kube-api-access-q764r" (OuterVolumeSpecName: "kube-api-access-q764r") pod "ff1f8f36-82ad-4771-b8de-b5dacf30a13a" (UID: "ff1f8f36-82ad-4771-b8de-b5dacf30a13a"). InnerVolumeSpecName "kube-api-access-q764r". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:22:04 crc kubenswrapper[5117]: I0321 00:22:04.741310 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-q764r\" (UniqueName: \"kubernetes.io/projected/ff1f8f36-82ad-4771-b8de-b5dacf30a13a-kube-api-access-q764r\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:05 crc kubenswrapper[5117]: I0321 00:22:05.320998 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567542-2fs89" event={"ID":"ff1f8f36-82ad-4771-b8de-b5dacf30a13a","Type":"ContainerDied","Data":"30fb760ff15f2ac4088149fee1df147c7c1a0725bf53e57ad293eb9250b3191f"} Mar 21 00:22:05 crc kubenswrapper[5117]: I0321 00:22:05.321049 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567542-2fs89" Mar 21 00:22:05 crc kubenswrapper[5117]: I0321 00:22:05.321063 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30fb760ff15f2ac4088149fee1df147c7c1a0725bf53e57ad293eb9250b3191f" Mar 21 00:22:05 crc kubenswrapper[5117]: I0321 00:22:05.380019 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567536-87psb"] Mar 21 00:22:05 crc kubenswrapper[5117]: I0321 00:22:05.383286 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567536-87psb"] Mar 21 00:22:05 crc kubenswrapper[5117]: I0321 00:22:05.587455 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="877a7600-33d3-446f-b83f-459c6a36cc69" path="/var/lib/kubelet/pods/877a7600-33d3-446f-b83f-459c6a36cc69/volumes" Mar 21 00:22:24 crc kubenswrapper[5117]: I0321 00:22:24.718840 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:22:24 crc kubenswrapper[5117]: I0321 00:22:24.719831 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:22:28 crc kubenswrapper[5117]: I0321 00:22:28.197384 5117 scope.go:117] "RemoveContainer" containerID="6d07328e753842f8bfde46d56a802751ad4f1670ad572ded18d108d02ec382d5" Mar 21 00:22:30 crc kubenswrapper[5117]: I0321 00:22:30.893929 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq"] Mar 21 00:22:30 crc kubenswrapper[5117]: I0321 00:22:30.894718 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" podUID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerName="kube-rbac-proxy" containerID="cri-o://f64a733df56c6d6741a0fd293d4e4fa09f79855c173e321da1ba6b1fc5e80390" gracePeriod=30 Mar 21 00:22:30 crc kubenswrapper[5117]: I0321 00:22:30.894933 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" podUID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerName="ovnkube-cluster-manager" containerID="cri-o://265ada799a3ab9199d0147cd05bd4d41a3acf3ba376cbef5fba0041b69d968cc" gracePeriod=30 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.128792 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9kggs"] Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.129450 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovn-controller" containerID="cri-o://d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda" gracePeriod=30 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.129929 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="sbdb" containerID="cri-o://0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063" gracePeriod=30 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.129987 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="nbdb" containerID="cri-o://590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041" gracePeriod=30 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.130032 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="northd" containerID="cri-o://66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b" gracePeriod=30 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.130071 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84" gracePeriod=30 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.130122 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kube-rbac-proxy-node" containerID="cri-o://87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717" gracePeriod=30 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.130170 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovn-acl-logging" containerID="cri-o://bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18" gracePeriod=30 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.161129 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.170548 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovnkube-controller" containerID="cri-o://f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23" gracePeriod=30 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.198607 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z"] Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.201122 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ff1f8f36-82ad-4771-b8de-b5dacf30a13a" containerName="oc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.201149 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1f8f36-82ad-4771-b8de-b5dacf30a13a" containerName="oc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.201169 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerName="ovnkube-cluster-manager" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.201179 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerName="ovnkube-cluster-manager" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.201193 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerName="kube-rbac-proxy" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.201199 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerName="kube-rbac-proxy" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.201358 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ff1f8f36-82ad-4771-b8de-b5dacf30a13a" containerName="oc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.201375 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerName="kube-rbac-proxy" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.201391 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerName="ovnkube-cluster-manager" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.203164 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcpq2\" (UniqueName: \"kubernetes.io/projected/1a068441-be86-4149-89e1-e1bb6bb0569b-kube-api-access-lcpq2\") pod \"1a068441-be86-4149-89e1-e1bb6bb0569b\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.203265 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-ovnkube-config\") pod \"1a068441-be86-4149-89e1-e1bb6bb0569b\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.203456 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1a068441-be86-4149-89e1-e1bb6bb0569b-ovn-control-plane-metrics-cert\") pod \"1a068441-be86-4149-89e1-e1bb6bb0569b\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.203664 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-env-overrides\") pod \"1a068441-be86-4149-89e1-e1bb6bb0569b\" (UID: \"1a068441-be86-4149-89e1-e1bb6bb0569b\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.204387 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1a068441-be86-4149-89e1-e1bb6bb0569b" (UID: "1a068441-be86-4149-89e1-e1bb6bb0569b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.204412 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1a068441-be86-4149-89e1-e1bb6bb0569b" (UID: "1a068441-be86-4149-89e1-e1bb6bb0569b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.210343 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.218482 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a068441-be86-4149-89e1-e1bb6bb0569b-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "1a068441-be86-4149-89e1-e1bb6bb0569b" (UID: "1a068441-be86-4149-89e1-e1bb6bb0569b"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.225237 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a068441-be86-4149-89e1-e1bb6bb0569b-kube-api-access-lcpq2" (OuterVolumeSpecName: "kube-api-access-lcpq2") pod "1a068441-be86-4149-89e1-e1bb6bb0569b" (UID: "1a068441-be86-4149-89e1-e1bb6bb0569b"). InnerVolumeSpecName "kube-api-access-lcpq2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.305132 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcvdt\" (UniqueName: \"kubernetes.io/projected/a219ef80-cf70-490a-ad4c-0d7c10bb367a-kube-api-access-mcvdt\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.305224 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a219ef80-cf70-490a-ad4c-0d7c10bb367a-env-overrides\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.305371 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a219ef80-cf70-490a-ad4c-0d7c10bb367a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.305471 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a219ef80-cf70-490a-ad4c-0d7c10bb367a-ovnkube-config\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.305539 5117 reconciler_common.go:299] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1a068441-be86-4149-89e1-e1bb6bb0569b-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.305568 5117 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.305587 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lcpq2\" (UniqueName: \"kubernetes.io/projected/1a068441-be86-4149-89e1-e1bb6bb0569b-kube-api-access-lcpq2\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.305604 5117 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1a068441-be86-4149-89e1-e1bb6bb0569b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.407012 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a219ef80-cf70-490a-ad4c-0d7c10bb367a-ovnkube-config\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.407624 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mcvdt\" (UniqueName: \"kubernetes.io/projected/a219ef80-cf70-490a-ad4c-0d7c10bb367a-kube-api-access-mcvdt\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.407783 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a219ef80-cf70-490a-ad4c-0d7c10bb367a-env-overrides\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.407939 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a219ef80-cf70-490a-ad4c-0d7c10bb367a-ovnkube-config\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.407949 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a219ef80-cf70-490a-ad4c-0d7c10bb367a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.408620 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a219ef80-cf70-490a-ad4c-0d7c10bb367a-env-overrides\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.414288 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a219ef80-cf70-490a-ad4c-0d7c10bb367a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.423814 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcvdt\" (UniqueName: \"kubernetes.io/projected/a219ef80-cf70-490a-ad4c-0d7c10bb367a-kube-api-access-mcvdt\") pod \"ovnkube-control-plane-97c9b6c48-52l6z\" (UID: \"a219ef80-cf70-490a-ad4c-0d7c10bb367a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.486296 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9kggs_ee6c0ddc-1c70-45c2-b50e-3cdce2f83288/ovn-acl-logging/0.log" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.486904 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9kggs_ee6c0ddc-1c70-45c2-b50e-3cdce2f83288/ovn-controller/0.log" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.487466 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509428 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-systemd\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509486 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-script-lib\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509568 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-netns\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509593 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmjsx\" (UniqueName: \"kubernetes.io/projected/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-kube-api-access-mmjsx\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509627 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-node-log\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509642 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-log-socket\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509664 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-systemd-units\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509684 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovn-node-metrics-cert\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509703 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-netd\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509718 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-env-overrides\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509752 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-ovn\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509785 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-slash\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509835 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-etc-openvswitch\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509852 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-bin\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509889 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509922 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-var-lib-openvswitch\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509942 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-openvswitch\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.509978 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-kubelet\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.510008 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-ovn-kubernetes\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.510058 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-config\") pod \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\" (UID: \"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288\") " Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.510702 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.510837 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-slash" (OuterVolumeSpecName: "host-slash") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.510884 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.510927 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.510968 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511020 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511060 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511100 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511141 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511188 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-log-socket" (OuterVolumeSpecName: "log-socket") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511226 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511331 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511381 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-node-log" (OuterVolumeSpecName: "node-log") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511417 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511620 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511550 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.511822 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.519485 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.524475 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-kube-api-access-mmjsx" (OuterVolumeSpecName: "kube-api-access-mmjsx") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "kube-api-access-mmjsx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.552036 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" (UID: "ee6c0ddc-1c70-45c2-b50e-3cdce2f83288"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.556731 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x56nc"] Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557549 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kube-rbac-proxy-node" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557576 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kube-rbac-proxy-node" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557592 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="nbdb" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557600 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="nbdb" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557616 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovnkube-controller" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557622 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovnkube-controller" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557643 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kube-rbac-proxy-ovn-metrics" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557652 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kube-rbac-proxy-ovn-metrics" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557665 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="northd" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557672 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="northd" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557699 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="sbdb" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557709 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="sbdb" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557725 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kubecfg-setup" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557733 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kubecfg-setup" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557742 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovn-controller" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557748 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovn-controller" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557761 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovn-acl-logging" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557769 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovn-acl-logging" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557894 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="northd" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557914 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovn-controller" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557926 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kube-rbac-proxy-ovn-metrics" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557934 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="sbdb" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557942 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovn-acl-logging" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557948 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="ovnkube-controller" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557958 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="nbdb" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.557966 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerName="kube-rbac-proxy-node" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.561050 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9kggs_ee6c0ddc-1c70-45c2-b50e-3cdce2f83288/ovn-acl-logging/0.log" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.561982 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9kggs_ee6c0ddc-1c70-45c2-b50e-3cdce2f83288/ovn-controller/0.log" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.562970 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerID="f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23" exitCode=0 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.563009 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerID="0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063" exitCode=0 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.563025 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerID="590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041" exitCode=0 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.563037 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerID="66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b" exitCode=0 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.563049 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerID="849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84" exitCode=0 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.563061 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerID="87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717" exitCode=0 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.563076 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerID="bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18" exitCode=143 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.563092 5117 generic.go:358] "Generic (PLEG): container finished" podID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" containerID="d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda" exitCode=143 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567792 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567857 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567871 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567881 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567892 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567918 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567931 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567944 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567950 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567957 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567969 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567976 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.567997 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568003 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568008 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568016 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568022 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568026 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568032 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568039 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568051 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568056 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568076 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568082 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568088 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568094 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568098 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568104 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568109 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568122 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" event={"ID":"ee6c0ddc-1c70-45c2-b50e-3cdce2f83288","Type":"ContainerDied","Data":"916c968b02e334a935d502ad04119e1d611ab3dfa2d51d7a7e47163721517311"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568132 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568156 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568163 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568169 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568175 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568181 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568186 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568192 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568196 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568196 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568073 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9kggs" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.568201 5117 scope.go:117] "RemoveContainer" containerID="f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.569533 5117 generic.go:358] "Generic (PLEG): container finished" podID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerID="265ada799a3ab9199d0147cd05bd4d41a3acf3ba376cbef5fba0041b69d968cc" exitCode=0 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.569649 5117 generic.go:358] "Generic (PLEG): container finished" podID="1a068441-be86-4149-89e1-e1bb6bb0569b" containerID="f64a733df56c6d6741a0fd293d4e4fa09f79855c173e321da1ba6b1fc5e80390" exitCode=0 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.569843 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" event={"ID":"1a068441-be86-4149-89e1-e1bb6bb0569b","Type":"ContainerDied","Data":"265ada799a3ab9199d0147cd05bd4d41a3acf3ba376cbef5fba0041b69d968cc"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.569953 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"265ada799a3ab9199d0147cd05bd4d41a3acf3ba376cbef5fba0041b69d968cc"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.570022 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f64a733df56c6d6741a0fd293d4e4fa09f79855c173e321da1ba6b1fc5e80390"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.570096 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" event={"ID":"1a068441-be86-4149-89e1-e1bb6bb0569b","Type":"ContainerDied","Data":"f64a733df56c6d6741a0fd293d4e4fa09f79855c173e321da1ba6b1fc5e80390"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.570172 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"265ada799a3ab9199d0147cd05bd4d41a3acf3ba376cbef5fba0041b69d968cc"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.570892 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f64a733df56c6d6741a0fd293d4e4fa09f79855c173e321da1ba6b1fc5e80390"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.570998 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" event={"ID":"1a068441-be86-4149-89e1-e1bb6bb0569b","Type":"ContainerDied","Data":"114386f51b93bcf86a51cb7dabf09cd927daa2b4b93cde30c17bade4a5bfbe87"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.571086 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"265ada799a3ab9199d0147cd05bd4d41a3acf3ba376cbef5fba0041b69d968cc"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.571241 5117 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f64a733df56c6d6741a0fd293d4e4fa09f79855c173e321da1ba6b1fc5e80390"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.570336 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.576548 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-72hnj_c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e/kube-multus/0.log" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.576646 5117 generic.go:358] "Generic (PLEG): container finished" podID="c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e" containerID="3f68e394d4ab5e40e8dc7c669538f180f3f69df9c5011076f34f1723b3472c4a" exitCode=2 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.576726 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-72hnj" event={"ID":"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e","Type":"ContainerDied","Data":"3f68e394d4ab5e40e8dc7c669538f180f3f69df9c5011076f34f1723b3472c4a"} Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.577960 5117 scope.go:117] "RemoveContainer" containerID="3f68e394d4ab5e40e8dc7c669538f180f3f69df9c5011076f34f1723b3472c4a" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.579418 5117 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.596637 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.611210 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfxrn\" (UniqueName: \"kubernetes.io/projected/c3fe16a8-bd52-427b-a473-cd852857427b-kube-api-access-cfxrn\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.611294 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-log-socket\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.611538 5117 scope.go:117] "RemoveContainer" containerID="0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.611750 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-run-systemd\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.611906 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-cni-bin\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.611958 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-var-lib-openvswitch\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.612033 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-kubelet\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.612201 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3fe16a8-bd52-427b-a473-cd852857427b-ovnkube-config\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.613474 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-run-ovn-kubernetes\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.613537 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-systemd-units\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.613577 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-run-openvswitch\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.613610 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.613647 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3fe16a8-bd52-427b-a473-cd852857427b-env-overrides\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.613847 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-run-ovn\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.613911 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-cni-netd\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.613941 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-etc-openvswitch\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.613997 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-run-netns\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614058 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3fe16a8-bd52-427b-a473-cd852857427b-ovnkube-script-lib\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614123 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-slash\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614166 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3fe16a8-bd52-427b-a473-cd852857427b-ovn-node-metrics-cert\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614229 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-node-log\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614590 5117 reconciler_common.go:299] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614619 5117 reconciler_common.go:299] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614641 5117 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614657 5117 reconciler_common.go:299] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614672 5117 reconciler_common.go:299] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614687 5117 reconciler_common.go:299] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614699 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mmjsx\" (UniqueName: \"kubernetes.io/projected/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-kube-api-access-mmjsx\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614712 5117 reconciler_common.go:299] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-node-log\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614724 5117 reconciler_common.go:299] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-log-socket\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614736 5117 reconciler_common.go:299] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614749 5117 reconciler_common.go:299] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614761 5117 reconciler_common.go:299] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614773 5117 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614786 5117 reconciler_common.go:299] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614797 5117 reconciler_common.go:299] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-slash\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614809 5117 reconciler_common.go:299] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614820 5117 reconciler_common.go:299] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614833 5117 reconciler_common.go:299] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614845 5117 reconciler_common.go:299] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.614861 5117 reconciler_common.go:299] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.639441 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq"] Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.644089 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-f4xkq"] Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.658763 5117 scope.go:117] "RemoveContainer" containerID="590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.681364 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9kggs"] Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.684552 5117 scope.go:117] "RemoveContainer" containerID="66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.686612 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9kggs"] Mar 21 00:22:31 crc kubenswrapper[5117]: W0321 00:22:31.693837 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda219ef80_cf70_490a_ad4c_0d7c10bb367a.slice/crio-6418fbf227830278eba0740ee01c19c8952f12f97bec16d6f92c4ddb661d2342 WatchSource:0}: Error finding container 6418fbf227830278eba0740ee01c19c8952f12f97bec16d6f92c4ddb661d2342: Status 404 returned error can't find the container with id 6418fbf227830278eba0740ee01c19c8952f12f97bec16d6f92c4ddb661d2342 Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.705008 5117 scope.go:117] "RemoveContainer" containerID="849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716424 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cfxrn\" (UniqueName: \"kubernetes.io/projected/c3fe16a8-bd52-427b-a473-cd852857427b-kube-api-access-cfxrn\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716523 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-log-socket\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716581 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-run-systemd\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716644 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-cni-bin\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716689 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-var-lib-openvswitch\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716725 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-kubelet\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716773 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3fe16a8-bd52-427b-a473-cd852857427b-ovnkube-config\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716887 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-run-ovn-kubernetes\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716935 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-systemd-units\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.716970 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-run-openvswitch\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717005 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717027 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-log-socket\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717037 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3fe16a8-bd52-427b-a473-cd852857427b-env-overrides\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717098 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-run-systemd\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717157 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-run-ovn\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717181 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-kubelet\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717188 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-cni-netd\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717228 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-cni-bin\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717267 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-run-ovn\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717279 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-etc-openvswitch\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717291 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-var-lib-openvswitch\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717335 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-run-netns\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717381 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3fe16a8-bd52-427b-a473-cd852857427b-ovnkube-script-lib\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717417 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-slash\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717450 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3fe16a8-bd52-427b-a473-cd852857427b-ovn-node-metrics-cert\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717487 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-node-log\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717587 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-node-log\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717637 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-run-openvswitch\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717699 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717959 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3fe16a8-bd52-427b-a473-cd852857427b-ovnkube-config\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.718007 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-run-ovn-kubernetes\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.718043 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-systemd-units\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.718091 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-run-netns\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.717208 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-cni-netd\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.718129 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-etc-openvswitch\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.718163 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3fe16a8-bd52-427b-a473-cd852857427b-host-slash\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.719016 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3fe16a8-bd52-427b-a473-cd852857427b-ovnkube-script-lib\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.719028 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3fe16a8-bd52-427b-a473-cd852857427b-env-overrides\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.727973 5117 scope.go:117] "RemoveContainer" containerID="87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.731373 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3fe16a8-bd52-427b-a473-cd852857427b-ovn-node-metrics-cert\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.736199 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfxrn\" (UniqueName: \"kubernetes.io/projected/c3fe16a8-bd52-427b-a473-cd852857427b-kube-api-access-cfxrn\") pod \"ovnkube-node-x56nc\" (UID: \"c3fe16a8-bd52-427b-a473-cd852857427b\") " pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.751036 5117 scope.go:117] "RemoveContainer" containerID="bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.783834 5117 scope.go:117] "RemoveContainer" containerID="d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.809591 5117 scope.go:117] "RemoveContainer" containerID="1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.831095 5117 scope.go:117] "RemoveContainer" containerID="f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23" Mar 21 00:22:31 crc kubenswrapper[5117]: E0321 00:22:31.831436 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": container with ID starting with f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23 not found: ID does not exist" containerID="f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.831492 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} err="failed to get container status \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": rpc error: code = NotFound desc = could not find container \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": container with ID starting with f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.831516 5117 scope.go:117] "RemoveContainer" containerID="0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063" Mar 21 00:22:31 crc kubenswrapper[5117]: E0321 00:22:31.833709 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": container with ID starting with 0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063 not found: ID does not exist" containerID="0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.833737 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} err="failed to get container status \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": rpc error: code = NotFound desc = could not find container \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": container with ID starting with 0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.833752 5117 scope.go:117] "RemoveContainer" containerID="590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041" Mar 21 00:22:31 crc kubenswrapper[5117]: E0321 00:22:31.834178 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": container with ID starting with 590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041 not found: ID does not exist" containerID="590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.834224 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} err="failed to get container status \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": rpc error: code = NotFound desc = could not find container \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": container with ID starting with 590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.834238 5117 scope.go:117] "RemoveContainer" containerID="66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b" Mar 21 00:22:31 crc kubenswrapper[5117]: E0321 00:22:31.834796 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": container with ID starting with 66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b not found: ID does not exist" containerID="66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.834821 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} err="failed to get container status \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": rpc error: code = NotFound desc = could not find container \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": container with ID starting with 66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.834840 5117 scope.go:117] "RemoveContainer" containerID="849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84" Mar 21 00:22:31 crc kubenswrapper[5117]: E0321 00:22:31.835322 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": container with ID starting with 849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84 not found: ID does not exist" containerID="849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.835363 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} err="failed to get container status \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": rpc error: code = NotFound desc = could not find container \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": container with ID starting with 849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.835380 5117 scope.go:117] "RemoveContainer" containerID="87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717" Mar 21 00:22:31 crc kubenswrapper[5117]: E0321 00:22:31.835910 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": container with ID starting with 87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717 not found: ID does not exist" containerID="87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.835952 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} err="failed to get container status \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": rpc error: code = NotFound desc = could not find container \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": container with ID starting with 87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.835967 5117 scope.go:117] "RemoveContainer" containerID="bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18" Mar 21 00:22:31 crc kubenswrapper[5117]: E0321 00:22:31.836328 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18\": container with ID starting with bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18 not found: ID does not exist" containerID="bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.836351 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} err="failed to get container status \"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18\": rpc error: code = NotFound desc = could not find container \"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18\": container with ID starting with bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.836363 5117 scope.go:117] "RemoveContainer" containerID="d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda" Mar 21 00:22:31 crc kubenswrapper[5117]: E0321 00:22:31.836594 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda\": container with ID starting with d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda not found: ID does not exist" containerID="d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.836612 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} err="failed to get container status \"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda\": rpc error: code = NotFound desc = could not find container \"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda\": container with ID starting with d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.836626 5117 scope.go:117] "RemoveContainer" containerID="1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34" Mar 21 00:22:31 crc kubenswrapper[5117]: E0321 00:22:31.836863 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\": container with ID starting with 1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34 not found: ID does not exist" containerID="1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.836890 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34"} err="failed to get container status \"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\": rpc error: code = NotFound desc = could not find container \"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\": container with ID starting with 1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.836902 5117 scope.go:117] "RemoveContainer" containerID="f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.837076 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} err="failed to get container status \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": rpc error: code = NotFound desc = could not find container \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": container with ID starting with f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.837103 5117 scope.go:117] "RemoveContainer" containerID="0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.838608 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} err="failed to get container status \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": rpc error: code = NotFound desc = could not find container \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": container with ID starting with 0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.838636 5117 scope.go:117] "RemoveContainer" containerID="590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.839157 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} err="failed to get container status \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": rpc error: code = NotFound desc = could not find container \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": container with ID starting with 590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.839184 5117 scope.go:117] "RemoveContainer" containerID="66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.839590 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} err="failed to get container status \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": rpc error: code = NotFound desc = could not find container \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": container with ID starting with 66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.839672 5117 scope.go:117] "RemoveContainer" containerID="849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.840437 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} err="failed to get container status \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": rpc error: code = NotFound desc = could not find container \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": container with ID starting with 849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.840460 5117 scope.go:117] "RemoveContainer" containerID="87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.844751 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} err="failed to get container status \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": rpc error: code = NotFound desc = could not find container \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": container with ID starting with 87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.844773 5117 scope.go:117] "RemoveContainer" containerID="bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.845066 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} err="failed to get container status \"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18\": rpc error: code = NotFound desc = could not find container \"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18\": container with ID starting with bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.845084 5117 scope.go:117] "RemoveContainer" containerID="d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.845615 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} err="failed to get container status \"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda\": rpc error: code = NotFound desc = could not find container \"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda\": container with ID starting with d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.845632 5117 scope.go:117] "RemoveContainer" containerID="1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.845987 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34"} err="failed to get container status \"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\": rpc error: code = NotFound desc = could not find container \"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\": container with ID starting with 1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.846012 5117 scope.go:117] "RemoveContainer" containerID="f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.846567 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} err="failed to get container status \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": rpc error: code = NotFound desc = could not find container \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": container with ID starting with f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.846623 5117 scope.go:117] "RemoveContainer" containerID="0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.847276 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} err="failed to get container status \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": rpc error: code = NotFound desc = could not find container \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": container with ID starting with 0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.847295 5117 scope.go:117] "RemoveContainer" containerID="590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.847718 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} err="failed to get container status \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": rpc error: code = NotFound desc = could not find container \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": container with ID starting with 590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.847760 5117 scope.go:117] "RemoveContainer" containerID="66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.848357 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} err="failed to get container status \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": rpc error: code = NotFound desc = could not find container \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": container with ID starting with 66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.848382 5117 scope.go:117] "RemoveContainer" containerID="849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.849822 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} err="failed to get container status \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": rpc error: code = NotFound desc = could not find container \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": container with ID starting with 849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.849850 5117 scope.go:117] "RemoveContainer" containerID="87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.850339 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} err="failed to get container status \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": rpc error: code = NotFound desc = could not find container \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": container with ID starting with 87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.850414 5117 scope.go:117] "RemoveContainer" containerID="bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.850853 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} err="failed to get container status \"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18\": rpc error: code = NotFound desc = could not find container \"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18\": container with ID starting with bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.850883 5117 scope.go:117] "RemoveContainer" containerID="d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.851171 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} err="failed to get container status \"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda\": rpc error: code = NotFound desc = could not find container \"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda\": container with ID starting with d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.851194 5117 scope.go:117] "RemoveContainer" containerID="1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.851854 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34"} err="failed to get container status \"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\": rpc error: code = NotFound desc = could not find container \"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\": container with ID starting with 1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.851878 5117 scope.go:117] "RemoveContainer" containerID="f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.852171 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} err="failed to get container status \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": rpc error: code = NotFound desc = could not find container \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": container with ID starting with f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.852226 5117 scope.go:117] "RemoveContainer" containerID="0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.852626 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} err="failed to get container status \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": rpc error: code = NotFound desc = could not find container \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": container with ID starting with 0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.852649 5117 scope.go:117] "RemoveContainer" containerID="590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.853007 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} err="failed to get container status \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": rpc error: code = NotFound desc = could not find container \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": container with ID starting with 590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.853028 5117 scope.go:117] "RemoveContainer" containerID="66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.853397 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} err="failed to get container status \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": rpc error: code = NotFound desc = could not find container \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": container with ID starting with 66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.853427 5117 scope.go:117] "RemoveContainer" containerID="849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.854442 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} err="failed to get container status \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": rpc error: code = NotFound desc = could not find container \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": container with ID starting with 849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.854461 5117 scope.go:117] "RemoveContainer" containerID="87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.854982 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} err="failed to get container status \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": rpc error: code = NotFound desc = could not find container \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": container with ID starting with 87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.854995 5117 scope.go:117] "RemoveContainer" containerID="bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.855418 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18"} err="failed to get container status \"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18\": rpc error: code = NotFound desc = could not find container \"bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18\": container with ID starting with bead42c9a05536558d78c8afb7648830cf183a4afd2c6c42c2006c745edd4d18 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.855440 5117 scope.go:117] "RemoveContainer" containerID="d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.855788 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda"} err="failed to get container status \"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda\": rpc error: code = NotFound desc = could not find container \"d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda\": container with ID starting with d30fa9143fddcc8ebaa12f88011fcf8b4a60cd11eaf3f161c6a82ca8bf2a6dda not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.855816 5117 scope.go:117] "RemoveContainer" containerID="1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.856076 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34"} err="failed to get container status \"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\": rpc error: code = NotFound desc = could not find container \"1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34\": container with ID starting with 1e476a97a6ef07899aa42f34f5e99e3e2daa02ff0ef1b453dc196c4092298f34 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.856098 5117 scope.go:117] "RemoveContainer" containerID="f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.856478 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23"} err="failed to get container status \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": rpc error: code = NotFound desc = could not find container \"f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23\": container with ID starting with f5edd82b52fd3e45684796c8d26f9ce07b5bbfebcbb9179ec5bef1466752ff23 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.856499 5117 scope.go:117] "RemoveContainer" containerID="0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.857034 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063"} err="failed to get container status \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": rpc error: code = NotFound desc = could not find container \"0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063\": container with ID starting with 0ef178f34690628847c3bf653da080a837791773e7f5c3bf3eba7729c1db1063 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.857056 5117 scope.go:117] "RemoveContainer" containerID="590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.857348 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041"} err="failed to get container status \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": rpc error: code = NotFound desc = could not find container \"590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041\": container with ID starting with 590bfc190bff683257ec4ac6cdab18a17ee58ece41b3c15b0b90a67ac8088041 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.857377 5117 scope.go:117] "RemoveContainer" containerID="66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.857809 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b"} err="failed to get container status \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": rpc error: code = NotFound desc = could not find container \"66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b\": container with ID starting with 66545dfd02090a6e210c450f447d8fc2bcc99bdc1baabb6e53a67ceb3e4a342b not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.857837 5117 scope.go:117] "RemoveContainer" containerID="849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.858518 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84"} err="failed to get container status \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": rpc error: code = NotFound desc = could not find container \"849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84\": container with ID starting with 849430cea498a775daeea3c3fcd9e4cc882422b054df7491a8dd8cfecf2c6c84 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.858542 5117 scope.go:117] "RemoveContainer" containerID="87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.867364 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717"} err="failed to get container status \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": rpc error: code = NotFound desc = could not find container \"87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717\": container with ID starting with 87b4571610f3a18ce19c52b56a03cf4a837c2b264a8894301a533c5e255a7717 not found: ID does not exist" Mar 21 00:22:31 crc kubenswrapper[5117]: I0321 00:22:31.896337 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:31 crc kubenswrapper[5117]: W0321 00:22:31.916552 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3fe16a8_bd52_427b_a473_cd852857427b.slice/crio-30d22271e6a3dcba4f009e20315a4f1357e7eaaf1b93a1c553aa1f9f41ab93da WatchSource:0}: Error finding container 30d22271e6a3dcba4f009e20315a4f1357e7eaaf1b93a1c553aa1f9f41ab93da: Status 404 returned error can't find the container with id 30d22271e6a3dcba4f009e20315a4f1357e7eaaf1b93a1c553aa1f9f41ab93da Mar 21 00:22:32 crc kubenswrapper[5117]: I0321 00:22:32.591446 5117 generic.go:358] "Generic (PLEG): container finished" podID="c3fe16a8-bd52-427b-a473-cd852857427b" containerID="7a87e56df839b43ce65968537bbc3c7458aad017642e797148c38970597c6a7b" exitCode=0 Mar 21 00:22:32 crc kubenswrapper[5117]: I0321 00:22:32.591526 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerDied","Data":"7a87e56df839b43ce65968537bbc3c7458aad017642e797148c38970597c6a7b"} Mar 21 00:22:32 crc kubenswrapper[5117]: I0321 00:22:32.591661 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerStarted","Data":"30d22271e6a3dcba4f009e20315a4f1357e7eaaf1b93a1c553aa1f9f41ab93da"} Mar 21 00:22:32 crc kubenswrapper[5117]: I0321 00:22:32.597754 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-72hnj_c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e/kube-multus/0.log" Mar 21 00:22:32 crc kubenswrapper[5117]: I0321 00:22:32.597954 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-72hnj" event={"ID":"c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e","Type":"ContainerStarted","Data":"cbaa956739bf83f49936b8ebec61ca5ddb813e8256f106afbd046b98b19554e2"} Mar 21 00:22:32 crc kubenswrapper[5117]: I0321 00:22:32.600781 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" event={"ID":"a219ef80-cf70-490a-ad4c-0d7c10bb367a","Type":"ContainerStarted","Data":"8f6f8da804f80a24b1617d99b99550a613a01f78c6f59eabfcc282e300d1664e"} Mar 21 00:22:32 crc kubenswrapper[5117]: I0321 00:22:32.600838 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" event={"ID":"a219ef80-cf70-490a-ad4c-0d7c10bb367a","Type":"ContainerStarted","Data":"8d42d1c94ccf5b7ab37748b30472144f7b071551c5984137d789e29bc9397a90"} Mar 21 00:22:32 crc kubenswrapper[5117]: I0321 00:22:32.600858 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" event={"ID":"a219ef80-cf70-490a-ad4c-0d7c10bb367a","Type":"ContainerStarted","Data":"6418fbf227830278eba0740ee01c19c8952f12f97bec16d6f92c4ddb661d2342"} Mar 21 00:22:32 crc kubenswrapper[5117]: I0321 00:22:32.657219 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-52l6z" podStartSLOduration=2.6571997830000003 podStartE2EDuration="2.657199783s" podCreationTimestamp="2026-03-21 00:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:22:32.653546369 +0000 UTC m=+665.946833591" watchObservedRunningTime="2026-03-21 00:22:32.657199783 +0000 UTC m=+665.950486945" Mar 21 00:22:33 crc kubenswrapper[5117]: I0321 00:22:33.594497 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a068441-be86-4149-89e1-e1bb6bb0569b" path="/var/lib/kubelet/pods/1a068441-be86-4149-89e1-e1bb6bb0569b/volumes" Mar 21 00:22:33 crc kubenswrapper[5117]: I0321 00:22:33.596280 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee6c0ddc-1c70-45c2-b50e-3cdce2f83288" path="/var/lib/kubelet/pods/ee6c0ddc-1c70-45c2-b50e-3cdce2f83288/volumes" Mar 21 00:22:33 crc kubenswrapper[5117]: I0321 00:22:33.614928 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerStarted","Data":"0c958bea5561494cc48770896c7b609818014e278c537dc88ecd275f561cd7d0"} Mar 21 00:22:33 crc kubenswrapper[5117]: I0321 00:22:33.615011 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerStarted","Data":"02ad37962835fcc3ad9e120806c5cc1a687b546cbdb252049891548e732bf95a"} Mar 21 00:22:33 crc kubenswrapper[5117]: I0321 00:22:33.615038 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerStarted","Data":"546d97f9cda616345e80ac8c3a8230c44f16be2033517d3dd5676a84f0cfc7a6"} Mar 21 00:22:33 crc kubenswrapper[5117]: I0321 00:22:33.615059 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerStarted","Data":"5836095e36e8f37ee9fd7cab0d141b5a9ab08114a1dac7c9d04341f8a7d94fa6"} Mar 21 00:22:33 crc kubenswrapper[5117]: I0321 00:22:33.615077 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerStarted","Data":"69e5603f066adda290b627e5f46631c309627c767f3afc5b8eedbf01990d2c8e"} Mar 21 00:22:34 crc kubenswrapper[5117]: I0321 00:22:34.628750 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerStarted","Data":"79762372fff8f6fcf499dd7637b281f13db11fd922674a697bdc7794e6d33e16"} Mar 21 00:22:36 crc kubenswrapper[5117]: I0321 00:22:36.654478 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerStarted","Data":"aca9693a68f670f33d7a064c7c56473c9a92e087a676a996e0a6145a2e475e0a"} Mar 21 00:22:38 crc kubenswrapper[5117]: I0321 00:22:38.692821 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" event={"ID":"c3fe16a8-bd52-427b-a473-cd852857427b","Type":"ContainerStarted","Data":"c48ae6cfa1014b483d3ec556f600b1c5cd5c3bcf51a9dccc399df347b5b91ab7"} Mar 21 00:22:38 crc kubenswrapper[5117]: I0321 00:22:38.693947 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:38 crc kubenswrapper[5117]: I0321 00:22:38.694140 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:38 crc kubenswrapper[5117]: I0321 00:22:38.694223 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:38 crc kubenswrapper[5117]: I0321 00:22:38.749899 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" podStartSLOduration=7.749873605 podStartE2EDuration="7.749873605s" podCreationTimestamp="2026-03-21 00:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:22:38.745913443 +0000 UTC m=+672.039200685" watchObservedRunningTime="2026-03-21 00:22:38.749873605 +0000 UTC m=+672.043160817" Mar 21 00:22:38 crc kubenswrapper[5117]: I0321 00:22:38.762534 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:38 crc kubenswrapper[5117]: I0321 00:22:38.769387 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:22:54 crc kubenswrapper[5117]: I0321 00:22:54.718157 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:22:54 crc kubenswrapper[5117]: I0321 00:22:54.718859 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:23:10 crc kubenswrapper[5117]: I0321 00:23:10.751184 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x56nc" Mar 21 00:23:24 crc kubenswrapper[5117]: I0321 00:23:24.718875 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:23:24 crc kubenswrapper[5117]: I0321 00:23:24.719721 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:23:24 crc kubenswrapper[5117]: I0321 00:23:24.719795 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:23:24 crc kubenswrapper[5117]: I0321 00:23:24.720703 5117 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2fd8729d8670a753e93005428f17279031e73d4ffabcffd4b6e8f29d6bcce8b"} pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 00:23:24 crc kubenswrapper[5117]: I0321 00:23:24.720810 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" containerID="cri-o://f2fd8729d8670a753e93005428f17279031e73d4ffabcffd4b6e8f29d6bcce8b" gracePeriod=600 Mar 21 00:23:25 crc kubenswrapper[5117]: I0321 00:23:25.080589 5117 generic.go:358] "Generic (PLEG): container finished" podID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerID="f2fd8729d8670a753e93005428f17279031e73d4ffabcffd4b6e8f29d6bcce8b" exitCode=0 Mar 21 00:23:25 crc kubenswrapper[5117]: I0321 00:23:25.080676 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerDied","Data":"f2fd8729d8670a753e93005428f17279031e73d4ffabcffd4b6e8f29d6bcce8b"} Mar 21 00:23:25 crc kubenswrapper[5117]: I0321 00:23:25.081168 5117 scope.go:117] "RemoveContainer" containerID="83c05081a1abbd89c7dc06294900b198f6c61bbe00ae5049cd1d47115a11e0ea" Mar 21 00:23:26 crc kubenswrapper[5117]: I0321 00:23:26.092051 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerStarted","Data":"324331d7c0a640c4e69a7d08e2bb3a2ebe5299eeda584b902a33658b2293b76e"} Mar 21 00:23:28 crc kubenswrapper[5117]: I0321 00:23:28.376171 5117 scope.go:117] "RemoveContainer" containerID="f64a733df56c6d6741a0fd293d4e4fa09f79855c173e321da1ba6b1fc5e80390" Mar 21 00:23:28 crc kubenswrapper[5117]: I0321 00:23:28.411365 5117 scope.go:117] "RemoveContainer" containerID="265ada799a3ab9199d0147cd05bd4d41a3acf3ba376cbef5fba0041b69d968cc" Mar 21 00:23:35 crc kubenswrapper[5117]: I0321 00:23:35.715470 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xr9n"] Mar 21 00:23:35 crc kubenswrapper[5117]: I0321 00:23:35.717337 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8xr9n" podUID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerName="registry-server" containerID="cri-o://baac535c8d9804bf9b83f22e50e9550ac413e25ff7243825ed49e4773dc6f84a" gracePeriod=30 Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.174723 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xr9n" event={"ID":"d91c7d7a-4c8e-454c-ae77-791050b6c6e3","Type":"ContainerDied","Data":"baac535c8d9804bf9b83f22e50e9550ac413e25ff7243825ed49e4773dc6f84a"} Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.174666 5117 generic.go:358] "Generic (PLEG): container finished" podID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerID="baac535c8d9804bf9b83f22e50e9550ac413e25ff7243825ed49e4773dc6f84a" exitCode=0 Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.175827 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8xr9n" event={"ID":"d91c7d7a-4c8e-454c-ae77-791050b6c6e3","Type":"ContainerDied","Data":"94c94d1e12072ce0b3ed8c3fbd500f7825120329d9a7648a73ff9490a0eb2ab1"} Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.175865 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94c94d1e12072ce0b3ed8c3fbd500f7825120329d9a7648a73ff9490a0eb2ab1" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.208708 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.348827 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-catalog-content\") pod \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.348962 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td66s\" (UniqueName: \"kubernetes.io/projected/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-kube-api-access-td66s\") pod \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.349089 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-utilities\") pod \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\" (UID: \"d91c7d7a-4c8e-454c-ae77-791050b6c6e3\") " Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.351959 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-utilities" (OuterVolumeSpecName: "utilities") pod "d91c7d7a-4c8e-454c-ae77-791050b6c6e3" (UID: "d91c7d7a-4c8e-454c-ae77-791050b6c6e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.361543 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-kube-api-access-td66s" (OuterVolumeSpecName: "kube-api-access-td66s") pod "d91c7d7a-4c8e-454c-ae77-791050b6c6e3" (UID: "d91c7d7a-4c8e-454c-ae77-791050b6c6e3"). InnerVolumeSpecName "kube-api-access-td66s". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.378280 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d91c7d7a-4c8e-454c-ae77-791050b6c6e3" (UID: "d91c7d7a-4c8e-454c-ae77-791050b6c6e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.450913 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.450982 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-td66s\" (UniqueName: \"kubernetes.io/projected/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-kube-api-access-td66s\") on node \"crc\" DevicePath \"\"" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.451007 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d91c7d7a-4c8e-454c-ae77-791050b6c6e3-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.676477 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-5d9d95bf5b-2qg8v"] Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.677429 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerName="registry-server" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.677463 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerName="registry-server" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.677531 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerName="extract-utilities" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.677551 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerName="extract-utilities" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.677570 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerName="extract-content" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.677591 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerName="extract-content" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.677855 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" containerName="registry-server" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.695801 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.698763 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-5d9d95bf5b-2qg8v"] Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.758483 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4gf6\" (UniqueName: \"kubernetes.io/projected/c49fcaa6-b88a-4614-af0c-04dba8e277af-kube-api-access-s4gf6\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.758562 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c49fcaa6-b88a-4614-af0c-04dba8e277af-trusted-ca\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.758596 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.758642 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c49fcaa6-b88a-4614-af0c-04dba8e277af-bound-sa-token\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.758681 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c49fcaa6-b88a-4614-af0c-04dba8e277af-registry-certificates\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.758741 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c49fcaa6-b88a-4614-af0c-04dba8e277af-installation-pull-secrets\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.758760 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c49fcaa6-b88a-4614-af0c-04dba8e277af-registry-tls\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.758779 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c49fcaa6-b88a-4614-af0c-04dba8e277af-ca-trust-extracted\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.835929 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.859749 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-s4gf6\" (UniqueName: \"kubernetes.io/projected/c49fcaa6-b88a-4614-af0c-04dba8e277af-kube-api-access-s4gf6\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.859811 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c49fcaa6-b88a-4614-af0c-04dba8e277af-trusted-ca\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.859844 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c49fcaa6-b88a-4614-af0c-04dba8e277af-bound-sa-token\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.859870 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c49fcaa6-b88a-4614-af0c-04dba8e277af-registry-certificates\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.859886 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c49fcaa6-b88a-4614-af0c-04dba8e277af-installation-pull-secrets\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.859904 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c49fcaa6-b88a-4614-af0c-04dba8e277af-registry-tls\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.859924 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c49fcaa6-b88a-4614-af0c-04dba8e277af-ca-trust-extracted\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.860337 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c49fcaa6-b88a-4614-af0c-04dba8e277af-ca-trust-extracted\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.861417 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c49fcaa6-b88a-4614-af0c-04dba8e277af-trusted-ca\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.862749 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c49fcaa6-b88a-4614-af0c-04dba8e277af-registry-certificates\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.867364 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c49fcaa6-b88a-4614-af0c-04dba8e277af-installation-pull-secrets\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.867380 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c49fcaa6-b88a-4614-af0c-04dba8e277af-registry-tls\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.879276 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4gf6\" (UniqueName: \"kubernetes.io/projected/c49fcaa6-b88a-4614-af0c-04dba8e277af-kube-api-access-s4gf6\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:36 crc kubenswrapper[5117]: I0321 00:23:36.885147 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c49fcaa6-b88a-4614-af0c-04dba8e277af-bound-sa-token\") pod \"image-registry-5d9d95bf5b-2qg8v\" (UID: \"c49fcaa6-b88a-4614-af0c-04dba8e277af\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:37 crc kubenswrapper[5117]: I0321 00:23:37.013587 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:37 crc kubenswrapper[5117]: I0321 00:23:37.185790 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8xr9n" Mar 21 00:23:37 crc kubenswrapper[5117]: I0321 00:23:37.243331 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xr9n"] Mar 21 00:23:37 crc kubenswrapper[5117]: I0321 00:23:37.248672 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8xr9n"] Mar 21 00:23:37 crc kubenswrapper[5117]: I0321 00:23:37.302133 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-5d9d95bf5b-2qg8v"] Mar 21 00:23:37 crc kubenswrapper[5117]: W0321 00:23:37.303217 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc49fcaa6_b88a_4614_af0c_04dba8e277af.slice/crio-588fd62a0b3a02dab8b93542580c6535d6b2c7c1118d477abe00af5746a4f4ad WatchSource:0}: Error finding container 588fd62a0b3a02dab8b93542580c6535d6b2c7c1118d477abe00af5746a4f4ad: Status 404 returned error can't find the container with id 588fd62a0b3a02dab8b93542580c6535d6b2c7c1118d477abe00af5746a4f4ad Mar 21 00:23:37 crc kubenswrapper[5117]: E0321 00:23:37.322756 5117 cadvisor_stats_provider.go:525] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd91c7d7a_4c8e_454c_ae77_791050b6c6e3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd91c7d7a_4c8e_454c_ae77_791050b6c6e3.slice/crio-94c94d1e12072ce0b3ed8c3fbd500f7825120329d9a7648a73ff9490a0eb2ab1\": RecentStats: unable to find data in memory cache]" Mar 21 00:23:37 crc kubenswrapper[5117]: I0321 00:23:37.589095 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d91c7d7a-4c8e-454c-ae77-791050b6c6e3" path="/var/lib/kubelet/pods/d91c7d7a-4c8e-454c-ae77-791050b6c6e3/volumes" Mar 21 00:23:38 crc kubenswrapper[5117]: I0321 00:23:38.191111 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" event={"ID":"c49fcaa6-b88a-4614-af0c-04dba8e277af","Type":"ContainerStarted","Data":"14298d15f08ce4547a76c8439d8ec762d46899ebe8e3a54cfde3b401537d78d1"} Mar 21 00:23:38 crc kubenswrapper[5117]: I0321 00:23:38.191194 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" event={"ID":"c49fcaa6-b88a-4614-af0c-04dba8e277af","Type":"ContainerStarted","Data":"588fd62a0b3a02dab8b93542580c6535d6b2c7c1118d477abe00af5746a4f4ad"} Mar 21 00:23:38 crc kubenswrapper[5117]: I0321 00:23:38.191305 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:38 crc kubenswrapper[5117]: I0321 00:23:38.233551 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" podStartSLOduration=2.233518869 podStartE2EDuration="2.233518869s" podCreationTimestamp="2026-03-21 00:23:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:23:38.214427689 +0000 UTC m=+731.507714851" watchObservedRunningTime="2026-03-21 00:23:38.233518869 +0000 UTC m=+731.526806041" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.393793 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh"] Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.407057 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.410866 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-b2ccr\"" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.416280 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh"] Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.500233 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.500330 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.500782 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwhtk\" (UniqueName: \"kubernetes.io/projected/dc83e427-aef5-4984-b9e0-72d49d75c064-kube-api-access-qwhtk\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.602622 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.602707 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.602949 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qwhtk\" (UniqueName: \"kubernetes.io/projected/dc83e427-aef5-4984-b9e0-72d49d75c064-kube-api-access-qwhtk\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.603390 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.603554 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.647387 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwhtk\" (UniqueName: \"kubernetes.io/projected/dc83e427-aef5-4984-b9e0-72d49d75c064-kube-api-access-qwhtk\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:39 crc kubenswrapper[5117]: I0321 00:23:39.751943 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:40 crc kubenswrapper[5117]: I0321 00:23:40.236887 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh"] Mar 21 00:23:41 crc kubenswrapper[5117]: I0321 00:23:41.218152 5117 generic.go:358] "Generic (PLEG): container finished" podID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerID="deef01b397cddab80cdc12791b4ef0630f4c8c95dfa9dccdca6892fa97d22f61" exitCode=0 Mar 21 00:23:41 crc kubenswrapper[5117]: I0321 00:23:41.218317 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" event={"ID":"dc83e427-aef5-4984-b9e0-72d49d75c064","Type":"ContainerDied","Data":"deef01b397cddab80cdc12791b4ef0630f4c8c95dfa9dccdca6892fa97d22f61"} Mar 21 00:23:41 crc kubenswrapper[5117]: I0321 00:23:41.218826 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" event={"ID":"dc83e427-aef5-4984-b9e0-72d49d75c064","Type":"ContainerStarted","Data":"f1807c19b6d4873ef1dabaddeef428aa98d1195bf9b67038caccdd8f7c53cd4f"} Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.568449 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-857q2"] Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.583232 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-857q2"] Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.583602 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.659342 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-utilities\") pod \"redhat-operators-857q2\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.659446 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slnb2\" (UniqueName: \"kubernetes.io/projected/35d47888-2755-479f-8c93-a27faae9a116-kube-api-access-slnb2\") pod \"redhat-operators-857q2\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.659511 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-catalog-content\") pod \"redhat-operators-857q2\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.760659 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-utilities\") pod \"redhat-operators-857q2\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.760739 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-slnb2\" (UniqueName: \"kubernetes.io/projected/35d47888-2755-479f-8c93-a27faae9a116-kube-api-access-slnb2\") pod \"redhat-operators-857q2\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.760819 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-catalog-content\") pod \"redhat-operators-857q2\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.761653 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-catalog-content\") pod \"redhat-operators-857q2\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.761697 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-utilities\") pod \"redhat-operators-857q2\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.797904 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-slnb2\" (UniqueName: \"kubernetes.io/projected/35d47888-2755-479f-8c93-a27faae9a116-kube-api-access-slnb2\") pod \"redhat-operators-857q2\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:42 crc kubenswrapper[5117]: I0321 00:23:42.928027 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:43 crc kubenswrapper[5117]: I0321 00:23:43.182082 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-857q2"] Mar 21 00:23:43 crc kubenswrapper[5117]: W0321 00:23:43.189835 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35d47888_2755_479f_8c93_a27faae9a116.slice/crio-2ebef2fbc31b4f38470a53dc5b5aaa13419b6cddad777f279c5dc4e456356900 WatchSource:0}: Error finding container 2ebef2fbc31b4f38470a53dc5b5aaa13419b6cddad777f279c5dc4e456356900: Status 404 returned error can't find the container with id 2ebef2fbc31b4f38470a53dc5b5aaa13419b6cddad777f279c5dc4e456356900 Mar 21 00:23:43 crc kubenswrapper[5117]: I0321 00:23:43.245889 5117 generic.go:358] "Generic (PLEG): container finished" podID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerID="98919b0431b2f5da6392c9519e272ff2720c99370fae3e9c6beb44019c21d0e2" exitCode=0 Mar 21 00:23:43 crc kubenswrapper[5117]: I0321 00:23:43.245955 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" event={"ID":"dc83e427-aef5-4984-b9e0-72d49d75c064","Type":"ContainerDied","Data":"98919b0431b2f5da6392c9519e272ff2720c99370fae3e9c6beb44019c21d0e2"} Mar 21 00:23:43 crc kubenswrapper[5117]: I0321 00:23:43.248912 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-857q2" event={"ID":"35d47888-2755-479f-8c93-a27faae9a116","Type":"ContainerStarted","Data":"2ebef2fbc31b4f38470a53dc5b5aaa13419b6cddad777f279c5dc4e456356900"} Mar 21 00:23:44 crc kubenswrapper[5117]: I0321 00:23:44.258889 5117 generic.go:358] "Generic (PLEG): container finished" podID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerID="5ceda86b6f089a7414ff46d67eff4100813ebfd9cb12667cd1d4d64f2e2f3cc3" exitCode=0 Mar 21 00:23:44 crc kubenswrapper[5117]: I0321 00:23:44.258969 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" event={"ID":"dc83e427-aef5-4984-b9e0-72d49d75c064","Type":"ContainerDied","Data":"5ceda86b6f089a7414ff46d67eff4100813ebfd9cb12667cd1d4d64f2e2f3cc3"} Mar 21 00:23:44 crc kubenswrapper[5117]: I0321 00:23:44.261352 5117 generic.go:358] "Generic (PLEG): container finished" podID="35d47888-2755-479f-8c93-a27faae9a116" containerID="549525b527c9afc6737077e66c4e3224f648e34eb607170b7dfafaf3ef200d41" exitCode=0 Mar 21 00:23:44 crc kubenswrapper[5117]: I0321 00:23:44.261457 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-857q2" event={"ID":"35d47888-2755-479f-8c93-a27faae9a116","Type":"ContainerDied","Data":"549525b527c9afc6737077e66c4e3224f648e34eb607170b7dfafaf3ef200d41"} Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.271341 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-857q2" event={"ID":"35d47888-2755-479f-8c93-a27faae9a116","Type":"ContainerStarted","Data":"510179a5ba913b4d5b4774f36f4dba7f18843b3ae7245403fd6c473ecb3370c4"} Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.536372 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.709438 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-bundle\") pod \"dc83e427-aef5-4984-b9e0-72d49d75c064\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.709599 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwhtk\" (UniqueName: \"kubernetes.io/projected/dc83e427-aef5-4984-b9e0-72d49d75c064-kube-api-access-qwhtk\") pod \"dc83e427-aef5-4984-b9e0-72d49d75c064\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.709691 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-util\") pod \"dc83e427-aef5-4984-b9e0-72d49d75c064\" (UID: \"dc83e427-aef5-4984-b9e0-72d49d75c064\") " Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.713698 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-bundle" (OuterVolumeSpecName: "bundle") pod "dc83e427-aef5-4984-b9e0-72d49d75c064" (UID: "dc83e427-aef5-4984-b9e0-72d49d75c064"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.718708 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc83e427-aef5-4984-b9e0-72d49d75c064-kube-api-access-qwhtk" (OuterVolumeSpecName: "kube-api-access-qwhtk") pod "dc83e427-aef5-4984-b9e0-72d49d75c064" (UID: "dc83e427-aef5-4984-b9e0-72d49d75c064"). InnerVolumeSpecName "kube-api-access-qwhtk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.731860 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-util" (OuterVolumeSpecName: "util") pod "dc83e427-aef5-4984-b9e0-72d49d75c064" (UID: "dc83e427-aef5-4984-b9e0-72d49d75c064"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.811558 5117 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.811595 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qwhtk\" (UniqueName: \"kubernetes.io/projected/dc83e427-aef5-4984-b9e0-72d49d75c064-kube-api-access-qwhtk\") on node \"crc\" DevicePath \"\"" Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.811609 5117 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc83e427-aef5-4984-b9e0-72d49d75c064-util\") on node \"crc\" DevicePath \"\"" Mar 21 00:23:45 crc kubenswrapper[5117]: I0321 00:23:45.999763 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q"] Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.000732 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerName="extract" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.000755 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerName="extract" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.000781 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerName="util" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.000793 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerName="util" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.000859 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerName="pull" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.000873 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerName="pull" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.001235 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="dc83e427-aef5-4984-b9e0-72d49d75c064" containerName="extract" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.008178 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.016470 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q"] Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.116246 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4vff\" (UniqueName: \"kubernetes.io/projected/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-kube-api-access-d4vff\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.116668 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.116924 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.218631 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d4vff\" (UniqueName: \"kubernetes.io/projected/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-kube-api-access-d4vff\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.218796 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.218887 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.219908 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.220319 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.252084 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4vff\" (UniqueName: \"kubernetes.io/projected/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-kube-api-access-d4vff\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.285988 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" event={"ID":"dc83e427-aef5-4984-b9e0-72d49d75c064","Type":"ContainerDied","Data":"f1807c19b6d4873ef1dabaddeef428aa98d1195bf9b67038caccdd8f7c53cd4f"} Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.286054 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.286064 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1807c19b6d4873ef1dabaddeef428aa98d1195bf9b67038caccdd8f7c53cd4f" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.288602 5117 generic.go:358] "Generic (PLEG): container finished" podID="35d47888-2755-479f-8c93-a27faae9a116" containerID="510179a5ba913b4d5b4774f36f4dba7f18843b3ae7245403fd6c473ecb3370c4" exitCode=0 Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.288701 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-857q2" event={"ID":"35d47888-2755-479f-8c93-a27faae9a116","Type":"ContainerDied","Data":"510179a5ba913b4d5b4774f36f4dba7f18843b3ae7245403fd6c473ecb3370c4"} Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.335523 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:46 crc kubenswrapper[5117]: I0321 00:23:46.543812 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q"] Mar 21 00:23:47 crc kubenswrapper[5117]: I0321 00:23:47.299542 5117 generic.go:358] "Generic (PLEG): container finished" podID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerID="3ac7f0c0d64a247a9541d8c56b8b32347dd7cf0d603524c7b1aff97688496f58" exitCode=0 Mar 21 00:23:47 crc kubenswrapper[5117]: I0321 00:23:47.300041 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" event={"ID":"1cc8a25c-81b3-4ddf-bf36-ccbf077722be","Type":"ContainerDied","Data":"3ac7f0c0d64a247a9541d8c56b8b32347dd7cf0d603524c7b1aff97688496f58"} Mar 21 00:23:47 crc kubenswrapper[5117]: I0321 00:23:47.300093 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" event={"ID":"1cc8a25c-81b3-4ddf-bf36-ccbf077722be","Type":"ContainerStarted","Data":"89c7f76dda811eb866d037f9cdcef107a7ed5a636ff032c568f0b43d7f5e7264"} Mar 21 00:23:47 crc kubenswrapper[5117]: I0321 00:23:47.306738 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-857q2" event={"ID":"35d47888-2755-479f-8c93-a27faae9a116","Type":"ContainerStarted","Data":"bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4"} Mar 21 00:23:47 crc kubenswrapper[5117]: I0321 00:23:47.367985 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-857q2" podStartSLOduration=4.632523828 podStartE2EDuration="5.3679557s" podCreationTimestamp="2026-03-21 00:23:42 +0000 UTC" firstStartedPulling="2026-03-21 00:23:44.262566943 +0000 UTC m=+737.555854155" lastFinishedPulling="2026-03-21 00:23:44.997998845 +0000 UTC m=+738.291286027" observedRunningTime="2026-03-21 00:23:47.367108924 +0000 UTC m=+740.660396136" watchObservedRunningTime="2026-03-21 00:23:47.3679557 +0000 UTC m=+740.661242902" Mar 21 00:23:48 crc kubenswrapper[5117]: I0321 00:23:48.318205 5117 generic.go:358] "Generic (PLEG): container finished" podID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerID="7dda293816f62a674bfd380850c385e27efac724f6973a85aea733d0c3eecf8c" exitCode=0 Mar 21 00:23:48 crc kubenswrapper[5117]: I0321 00:23:48.318330 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" event={"ID":"1cc8a25c-81b3-4ddf-bf36-ccbf077722be","Type":"ContainerDied","Data":"7dda293816f62a674bfd380850c385e27efac724f6973a85aea733d0c3eecf8c"} Mar 21 00:23:49 crc kubenswrapper[5117]: I0321 00:23:49.332803 5117 generic.go:358] "Generic (PLEG): container finished" podID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerID="f72261dc369054c39f8ce2d7d95c9dcf505d23171b132b1a4753f55e329f6879" exitCode=0 Mar 21 00:23:49 crc kubenswrapper[5117]: I0321 00:23:49.333361 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" event={"ID":"1cc8a25c-81b3-4ddf-bf36-ccbf077722be","Type":"ContainerDied","Data":"f72261dc369054c39f8ce2d7d95c9dcf505d23171b132b1a4753f55e329f6879"} Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.556126 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bqhkb"] Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.616642 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bqhkb"] Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.616822 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.660447 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.694931 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-util\") pod \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.695023 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-bundle\") pod \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.695047 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4vff\" (UniqueName: \"kubernetes.io/projected/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-kube-api-access-d4vff\") pod \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\" (UID: \"1cc8a25c-81b3-4ddf-bf36-ccbf077722be\") " Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.695321 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-utilities\") pod \"certified-operators-bqhkb\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.695381 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-catalog-content\") pod \"certified-operators-bqhkb\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.695441 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bncw5\" (UniqueName: \"kubernetes.io/projected/d9d70414-581a-4fa9-89c6-54362e2659d5-kube-api-access-bncw5\") pod \"certified-operators-bqhkb\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.696156 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-bundle" (OuterVolumeSpecName: "bundle") pod "1cc8a25c-81b3-4ddf-bf36-ccbf077722be" (UID: "1cc8a25c-81b3-4ddf-bf36-ccbf077722be"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.713694 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-util" (OuterVolumeSpecName: "util") pod "1cc8a25c-81b3-4ddf-bf36-ccbf077722be" (UID: "1cc8a25c-81b3-4ddf-bf36-ccbf077722be"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.725884 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-kube-api-access-d4vff" (OuterVolumeSpecName: "kube-api-access-d4vff") pod "1cc8a25c-81b3-4ddf-bf36-ccbf077722be" (UID: "1cc8a25c-81b3-4ddf-bf36-ccbf077722be"). InnerVolumeSpecName "kube-api-access-d4vff". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.796352 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bncw5\" (UniqueName: \"kubernetes.io/projected/d9d70414-581a-4fa9-89c6-54362e2659d5-kube-api-access-bncw5\") pod \"certified-operators-bqhkb\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.796443 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-utilities\") pod \"certified-operators-bqhkb\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.796481 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-catalog-content\") pod \"certified-operators-bqhkb\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.796549 5117 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-util\") on node \"crc\" DevicePath \"\"" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.796562 5117 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.796574 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d4vff\" (UniqueName: \"kubernetes.io/projected/1cc8a25c-81b3-4ddf-bf36-ccbf077722be-kube-api-access-d4vff\") on node \"crc\" DevicePath \"\"" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.796987 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-utilities\") pod \"certified-operators-bqhkb\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.799283 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-catalog-content\") pod \"certified-operators-bqhkb\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.835040 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bncw5\" (UniqueName: \"kubernetes.io/projected/d9d70414-581a-4fa9-89c6-54362e2659d5-kube-api-access-bncw5\") pod \"certified-operators-bqhkb\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:50 crc kubenswrapper[5117]: I0321 00:23:50.971341 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:23:51 crc kubenswrapper[5117]: I0321 00:23:51.355054 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" Mar 21 00:23:51 crc kubenswrapper[5117]: I0321 00:23:51.355054 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q" event={"ID":"1cc8a25c-81b3-4ddf-bf36-ccbf077722be","Type":"ContainerDied","Data":"89c7f76dda811eb866d037f9cdcef107a7ed5a636ff032c568f0b43d7f5e7264"} Mar 21 00:23:51 crc kubenswrapper[5117]: I0321 00:23:51.355590 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89c7f76dda811eb866d037f9cdcef107a7ed5a636ff032c568f0b43d7f5e7264" Mar 21 00:23:51 crc kubenswrapper[5117]: I0321 00:23:51.463148 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bqhkb"] Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.361985 5117 generic.go:358] "Generic (PLEG): container finished" podID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerID="7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86" exitCode=0 Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.362051 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqhkb" event={"ID":"d9d70414-581a-4fa9-89c6-54362e2659d5","Type":"ContainerDied","Data":"7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86"} Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.362514 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqhkb" event={"ID":"d9d70414-581a-4fa9-89c6-54362e2659d5","Type":"ContainerStarted","Data":"fb291d257f6efb5ee572defa428aa5162acaafe4bbeebb2dcdf3cc2c8b22bbf0"} Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.483405 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st"] Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.483942 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerName="pull" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.483958 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerName="pull" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.483986 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerName="util" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.483992 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerName="util" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.484006 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerName="extract" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.484011 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerName="extract" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.484106 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="1cc8a25c-81b3-4ddf-bf36-ccbf077722be" containerName="extract" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.489127 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.491120 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-b2ccr\"" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.498170 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st"] Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.519315 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqvlq\" (UniqueName: \"kubernetes.io/projected/79ef4008-b3a9-4f54-93b2-c296b982c5e7-kube-api-access-nqvlq\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.519450 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.519795 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.621362 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.621579 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.621611 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nqvlq\" (UniqueName: \"kubernetes.io/projected/79ef4008-b3a9-4f54-93b2-c296b982c5e7-kube-api-access-nqvlq\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.621938 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.622035 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.652455 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqvlq\" (UniqueName: \"kubernetes.io/projected/79ef4008-b3a9-4f54-93b2-c296b982c5e7-kube-api-access-nqvlq\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.803725 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.931797 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:52 crc kubenswrapper[5117]: I0321 00:23:52.931898 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:23:53 crc kubenswrapper[5117]: I0321 00:23:53.183011 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st"] Mar 21 00:23:53 crc kubenswrapper[5117]: I0321 00:23:53.370315 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqhkb" event={"ID":"d9d70414-581a-4fa9-89c6-54362e2659d5","Type":"ContainerStarted","Data":"3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006"} Mar 21 00:23:53 crc kubenswrapper[5117]: I0321 00:23:53.377278 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" event={"ID":"79ef4008-b3a9-4f54-93b2-c296b982c5e7","Type":"ContainerStarted","Data":"b5037cb18c9338f6fc8a5e8437f5914300039004df97d8be728dd98069dfa975"} Mar 21 00:23:53 crc kubenswrapper[5117]: I0321 00:23:53.377719 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" event={"ID":"79ef4008-b3a9-4f54-93b2-c296b982c5e7","Type":"ContainerStarted","Data":"382c248d7f48dde9999c14e22b6dcbd0a0c90f8658d1d1d7aae2188f64f1c4aa"} Mar 21 00:23:53 crc kubenswrapper[5117]: I0321 00:23:53.944142 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zqp24"] Mar 21 00:23:53 crc kubenswrapper[5117]: I0321 00:23:53.948630 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:53 crc kubenswrapper[5117]: I0321 00:23:53.956162 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqp24"] Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.017389 5117 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-857q2" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="registry-server" probeResult="failure" output=< Mar 21 00:23:54 crc kubenswrapper[5117]: timeout: failed to connect service ":50051" within 1s Mar 21 00:23:54 crc kubenswrapper[5117]: > Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.040952 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-catalog-content\") pod \"community-operators-zqp24\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.041011 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-utilities\") pod \"community-operators-zqp24\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.041043 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdzgd\" (UniqueName: \"kubernetes.io/projected/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-kube-api-access-tdzgd\") pod \"community-operators-zqp24\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.142415 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-utilities\") pod \"community-operators-zqp24\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.142486 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-tdzgd\" (UniqueName: \"kubernetes.io/projected/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-kube-api-access-tdzgd\") pod \"community-operators-zqp24\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.142603 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-catalog-content\") pod \"community-operators-zqp24\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.143182 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-catalog-content\") pod \"community-operators-zqp24\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.143504 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-utilities\") pod \"community-operators-zqp24\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.171575 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdzgd\" (UniqueName: \"kubernetes.io/projected/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-kube-api-access-tdzgd\") pod \"community-operators-zqp24\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.265456 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.415690 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" event={"ID":"79ef4008-b3a9-4f54-93b2-c296b982c5e7","Type":"ContainerDied","Data":"b5037cb18c9338f6fc8a5e8437f5914300039004df97d8be728dd98069dfa975"} Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.415859 5117 generic.go:358] "Generic (PLEG): container finished" podID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerID="b5037cb18c9338f6fc8a5e8437f5914300039004df97d8be728dd98069dfa975" exitCode=0 Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.426899 5117 generic.go:358] "Generic (PLEG): container finished" podID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerID="3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006" exitCode=0 Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.426984 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqhkb" event={"ID":"d9d70414-581a-4fa9-89c6-54362e2659d5","Type":"ContainerDied","Data":"3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006"} Mar 21 00:23:54 crc kubenswrapper[5117]: I0321 00:23:54.632650 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqp24"] Mar 21 00:23:54 crc kubenswrapper[5117]: W0321 00:23:54.635034 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2dd6b73b_a33f_4bd3_8702_afe3e1a5542c.slice/crio-bdb87cd77dbaee610a1e19ca9d4dceb4f271cb7beb0436c4874538b2ec41a6e7 WatchSource:0}: Error finding container bdb87cd77dbaee610a1e19ca9d4dceb4f271cb7beb0436c4874538b2ec41a6e7: Status 404 returned error can't find the container with id bdb87cd77dbaee610a1e19ca9d4dceb4f271cb7beb0436c4874538b2ec41a6e7 Mar 21 00:23:55 crc kubenswrapper[5117]: I0321 00:23:55.437656 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqhkb" event={"ID":"d9d70414-581a-4fa9-89c6-54362e2659d5","Type":"ContainerStarted","Data":"1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3"} Mar 21 00:23:55 crc kubenswrapper[5117]: I0321 00:23:55.439552 5117 generic.go:358] "Generic (PLEG): container finished" podID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerID="c518f84d824d986729678ae3f1381ba60951ffaff9dfe8a05b195fe17fcaad4c" exitCode=0 Mar 21 00:23:55 crc kubenswrapper[5117]: I0321 00:23:55.439641 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqp24" event={"ID":"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c","Type":"ContainerDied","Data":"c518f84d824d986729678ae3f1381ba60951ffaff9dfe8a05b195fe17fcaad4c"} Mar 21 00:23:55 crc kubenswrapper[5117]: I0321 00:23:55.439671 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqp24" event={"ID":"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c","Type":"ContainerStarted","Data":"bdb87cd77dbaee610a1e19ca9d4dceb4f271cb7beb0436c4874538b2ec41a6e7"} Mar 21 00:23:55 crc kubenswrapper[5117]: I0321 00:23:55.472432 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bqhkb" podStartSLOduration=4.9161249399999996 podStartE2EDuration="5.472416992s" podCreationTimestamp="2026-03-21 00:23:50 +0000 UTC" firstStartedPulling="2026-03-21 00:23:52.362871076 +0000 UTC m=+745.656158248" lastFinishedPulling="2026-03-21 00:23:52.919163128 +0000 UTC m=+746.212450300" observedRunningTime="2026-03-21 00:23:55.469548133 +0000 UTC m=+748.762835305" watchObservedRunningTime="2026-03-21 00:23:55.472416992 +0000 UTC m=+748.765704164" Mar 21 00:23:57 crc kubenswrapper[5117]: I0321 00:23:57.454724 5117 generic.go:358] "Generic (PLEG): container finished" podID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerID="ba1ae419a9a827a323a81e420056b68ddac1648f94a4a35840d8a25dcdd0c515" exitCode=0 Mar 21 00:23:57 crc kubenswrapper[5117]: I0321 00:23:57.454991 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqp24" event={"ID":"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c","Type":"ContainerDied","Data":"ba1ae419a9a827a323a81e420056b68ddac1648f94a4a35840d8a25dcdd0c515"} Mar 21 00:23:58 crc kubenswrapper[5117]: I0321 00:23:58.849083 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl"] Mar 21 00:23:58 crc kubenswrapper[5117]: I0321 00:23:58.900078 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl"] Mar 21 00:23:58 crc kubenswrapper[5117]: I0321 00:23:58.900378 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl" Mar 21 00:23:58 crc kubenswrapper[5117]: I0321 00:23:58.903420 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"obo-prometheus-operator-dockercfg-bvztk\"" Mar 21 00:23:58 crc kubenswrapper[5117]: I0321 00:23:58.904069 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operators\"/\"openshift-service-ca.crt\"" Mar 21 00:23:58 crc kubenswrapper[5117]: I0321 00:23:58.904221 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operators\"/\"kube-root-ca.crt\"" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.014854 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2vhd\" (UniqueName: \"kubernetes.io/projected/5f884179-adb4-4b62-bec6-14f40db13be5-kube-api-access-r2vhd\") pod \"obo-prometheus-operator-55568fc96c-nv4sl\" (UID: \"5f884179-adb4-4b62-bec6-14f40db13be5\") " pod="openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.115676 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-r2vhd\" (UniqueName: \"kubernetes.io/projected/5f884179-adb4-4b62-bec6-14f40db13be5-kube-api-access-r2vhd\") pod \"obo-prometheus-operator-55568fc96c-nv4sl\" (UID: \"5f884179-adb4-4b62-bec6-14f40db13be5\") " pod="openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.135554 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2vhd\" (UniqueName: \"kubernetes.io/projected/5f884179-adb4-4b62-bec6-14f40db13be5-kube-api-access-r2vhd\") pod \"obo-prometheus-operator-55568fc96c-nv4sl\" (UID: \"5f884179-adb4-4b62-bec6-14f40db13be5\") " pod="openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.219465 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-5d9d95bf5b-2qg8v" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.219660 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.322768 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-gjth4"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.401831 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.429167 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.430193 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.432668 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.434842 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"obo-prometheus-operator-admission-webhook-dockercfg-lrwwz\"" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.435389 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"obo-prometheus-operator-admission-webhook-service-cert\"" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.438578 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.452011 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.477576 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" event={"ID":"79ef4008-b3a9-4f54-93b2-c296b982c5e7","Type":"ContainerStarted","Data":"435ea4f262366c63d34ce3eb98403b88d5cffbb4deed5a20e239a3b28a4a95a8"} Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.492572 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqp24" event={"ID":"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c","Type":"ContainerStarted","Data":"2575d19b634e688e0a9fbe0be767be50c2975341463d0348abbc7b7b1cfeb5f4"} Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.529344 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47f630cc-5d71-4431-8dea-397bfd008097-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-mggcv\" (UID: \"47f630cc-5d71-4431-8dea-397bfd008097\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.529395 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/68bdcac8-762b-4e94-8494-b571a8622dd8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-pvrmj\" (UID: \"68bdcac8-762b-4e94-8494-b571a8622dd8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.529503 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/68bdcac8-762b-4e94-8494-b571a8622dd8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-pvrmj\" (UID: \"68bdcac8-762b-4e94-8494-b571a8622dd8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.529535 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47f630cc-5d71-4431-8dea-397bfd008097-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-mggcv\" (UID: \"47f630cc-5d71-4431-8dea-397bfd008097\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.631360 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47f630cc-5d71-4431-8dea-397bfd008097-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-mggcv\" (UID: \"47f630cc-5d71-4431-8dea-397bfd008097\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.631415 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47f630cc-5d71-4431-8dea-397bfd008097-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-mggcv\" (UID: \"47f630cc-5d71-4431-8dea-397bfd008097\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.631443 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/68bdcac8-762b-4e94-8494-b571a8622dd8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-pvrmj\" (UID: \"68bdcac8-762b-4e94-8494-b571a8622dd8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.631515 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/68bdcac8-762b-4e94-8494-b571a8622dd8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-pvrmj\" (UID: \"68bdcac8-762b-4e94-8494-b571a8622dd8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.641167 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47f630cc-5d71-4431-8dea-397bfd008097-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-mggcv\" (UID: \"47f630cc-5d71-4431-8dea-397bfd008097\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.643317 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/68bdcac8-762b-4e94-8494-b571a8622dd8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-pvrmj\" (UID: \"68bdcac8-762b-4e94-8494-b571a8622dd8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.646560 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/68bdcac8-762b-4e94-8494-b571a8622dd8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-pvrmj\" (UID: \"68bdcac8-762b-4e94-8494-b571a8622dd8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.646915 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47f630cc-5d71-4431-8dea-397bfd008097-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5866674d46-mggcv\" (UID: \"47f630cc-5d71-4431-8dea-397bfd008097\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.738977 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zqp24" podStartSLOduration=5.874778383 podStartE2EDuration="6.738959605s" podCreationTimestamp="2026-03-21 00:23:53 +0000 UTC" firstStartedPulling="2026-03-21 00:23:55.440427003 +0000 UTC m=+748.733714175" lastFinishedPulling="2026-03-21 00:23:56.304608225 +0000 UTC m=+749.597895397" observedRunningTime="2026-03-21 00:23:59.523631517 +0000 UTC m=+752.816918689" watchObservedRunningTime="2026-03-21 00:23:59.738959605 +0000 UTC m=+753.032246777" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.741943 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-587f9c8867-6xfv5"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.747407 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.749612 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"observability-operator-sa-dockercfg-4hppn\"" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.751571 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.754794 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"observability-operator-tls\"" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.768089 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-587f9c8867-6xfv5"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.773430 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.821074 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.835542 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/61a55f96-3dbb-43c7-9749-71d00e304e57-observability-operator-tls\") pod \"observability-operator-587f9c8867-6xfv5\" (UID: \"61a55f96-3dbb-43c7-9749-71d00e304e57\") " pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.835603 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxjcb\" (UniqueName: \"kubernetes.io/projected/61a55f96-3dbb-43c7-9749-71d00e304e57-kube-api-access-lxjcb\") pod \"observability-operator-587f9c8867-6xfv5\" (UID: \"61a55f96-3dbb-43c7-9749-71d00e304e57\") " pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:23:59 crc kubenswrapper[5117]: W0321 00:23:59.846683 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f884179_adb4_4b62_bec6_14f40db13be5.slice/crio-45807791dbb5f37ba01f581ed4f0619f70c84851038525edf83842cb02b1ac7e WatchSource:0}: Error finding container 45807791dbb5f37ba01f581ed4f0619f70c84851038525edf83842cb02b1ac7e: Status 404 returned error can't find the container with id 45807791dbb5f37ba01f581ed4f0619f70c84851038525edf83842cb02b1ac7e Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.937110 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/61a55f96-3dbb-43c7-9749-71d00e304e57-observability-operator-tls\") pod \"observability-operator-587f9c8867-6xfv5\" (UID: \"61a55f96-3dbb-43c7-9749-71d00e304e57\") " pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.937185 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lxjcb\" (UniqueName: \"kubernetes.io/projected/61a55f96-3dbb-43c7-9749-71d00e304e57-kube-api-access-lxjcb\") pod \"observability-operator-587f9c8867-6xfv5\" (UID: \"61a55f96-3dbb-43c7-9749-71d00e304e57\") " pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.958584 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/61a55f96-3dbb-43c7-9749-71d00e304e57-observability-operator-tls\") pod \"observability-operator-587f9c8867-6xfv5\" (UID: \"61a55f96-3dbb-43c7-9749-71d00e304e57\") " pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.962431 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-7d56df5895-sgkqd"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.978993 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.980625 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-7d56df5895-sgkqd"] Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.981968 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxjcb\" (UniqueName: \"kubernetes.io/projected/61a55f96-3dbb-43c7-9749-71d00e304e57-kube-api-access-lxjcb\") pod \"observability-operator-587f9c8867-6xfv5\" (UID: \"61a55f96-3dbb-43c7-9749-71d00e304e57\") " pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.986437 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elastic-operator-service-cert\"" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.986642 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elastic-operator-dockercfg-fz7fd\"" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.986788 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"openshift-service-ca.crt\"" Mar 21 00:23:59 crc kubenswrapper[5117]: I0321 00:23:59.986925 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"kube-root-ca.crt\"" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.025310 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj"] Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.038927 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfkts\" (UniqueName: \"kubernetes.io/projected/0025749a-4a5d-4ee1-8de3-9dfe2093085d-kube-api-access-xfkts\") pod \"elastic-operator-7d56df5895-sgkqd\" (UID: \"0025749a-4a5d-4ee1-8de3-9dfe2093085d\") " pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.039001 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0025749a-4a5d-4ee1-8de3-9dfe2093085d-apiservice-cert\") pod \"elastic-operator-7d56df5895-sgkqd\" (UID: \"0025749a-4a5d-4ee1-8de3-9dfe2093085d\") " pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.039040 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0025749a-4a5d-4ee1-8de3-9dfe2093085d-webhook-cert\") pod \"elastic-operator-7d56df5895-sgkqd\" (UID: \"0025749a-4a5d-4ee1-8de3-9dfe2093085d\") " pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.070421 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.155085 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0025749a-4a5d-4ee1-8de3-9dfe2093085d-apiservice-cert\") pod \"elastic-operator-7d56df5895-sgkqd\" (UID: \"0025749a-4a5d-4ee1-8de3-9dfe2093085d\") " pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.155427 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0025749a-4a5d-4ee1-8de3-9dfe2093085d-webhook-cert\") pod \"elastic-operator-7d56df5895-sgkqd\" (UID: \"0025749a-4a5d-4ee1-8de3-9dfe2093085d\") " pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.155484 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xfkts\" (UniqueName: \"kubernetes.io/projected/0025749a-4a5d-4ee1-8de3-9dfe2093085d-kube-api-access-xfkts\") pod \"elastic-operator-7d56df5895-sgkqd\" (UID: \"0025749a-4a5d-4ee1-8de3-9dfe2093085d\") " pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.163174 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567544-rmr2v"] Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.165887 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0025749a-4a5d-4ee1-8de3-9dfe2093085d-apiservice-cert\") pod \"elastic-operator-7d56df5895-sgkqd\" (UID: \"0025749a-4a5d-4ee1-8de3-9dfe2093085d\") " pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.168852 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0025749a-4a5d-4ee1-8de3-9dfe2093085d-webhook-cert\") pod \"elastic-operator-7d56df5895-sgkqd\" (UID: \"0025749a-4a5d-4ee1-8de3-9dfe2093085d\") " pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.183267 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567544-rmr2v"] Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.183408 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567544-rmr2v" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.188614 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.188820 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.188970 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.215127 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfkts\" (UniqueName: \"kubernetes.io/projected/0025749a-4a5d-4ee1-8de3-9dfe2093085d-kube-api-access-xfkts\") pod \"elastic-operator-7d56df5895-sgkqd\" (UID: \"0025749a-4a5d-4ee1-8de3-9dfe2093085d\") " pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.258412 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44tcb\" (UniqueName: \"kubernetes.io/projected/be0be9be-183f-44ae-bea1-13b4c356bd8c-kube-api-access-44tcb\") pod \"auto-csr-approver-29567544-rmr2v\" (UID: \"be0be9be-183f-44ae-bea1-13b4c356bd8c\") " pod="openshift-infra/auto-csr-approver-29567544-rmr2v" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.285966 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv"] Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.314126 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.360152 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-44tcb\" (UniqueName: \"kubernetes.io/projected/be0be9be-183f-44ae-bea1-13b4c356bd8c-kube-api-access-44tcb\") pod \"auto-csr-approver-29567544-rmr2v\" (UID: \"be0be9be-183f-44ae-bea1-13b4c356bd8c\") " pod="openshift-infra/auto-csr-approver-29567544-rmr2v" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.384891 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-44tcb\" (UniqueName: \"kubernetes.io/projected/be0be9be-183f-44ae-bea1-13b4c356bd8c-kube-api-access-44tcb\") pod \"auto-csr-approver-29567544-rmr2v\" (UID: \"be0be9be-183f-44ae-bea1-13b4c356bd8c\") " pod="openshift-infra/auto-csr-approver-29567544-rmr2v" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.388744 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-b577b4f9-4xhsq"] Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.405469 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-b577b4f9-4xhsq"] Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.405648 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.409736 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"perses-operator-service-cert\"" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.409969 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"perses-operator-dockercfg-wgznh\"" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.463600 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e836407-e9f1-4125-9810-741ae3b23431-webhook-cert\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.463654 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9zh5\" (UniqueName: \"kubernetes.io/projected/4e836407-e9f1-4125-9810-741ae3b23431-kube-api-access-m9zh5\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.463712 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e836407-e9f1-4125-9810-741ae3b23431-apiservice-cert\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.463975 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/4e836407-e9f1-4125-9810-741ae3b23431-openshift-service-ca\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.507056 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" event={"ID":"47f630cc-5d71-4431-8dea-397bfd008097","Type":"ContainerStarted","Data":"9be13843d7472cd9daf81db82f9b7cc2892253bcfd5811d00e1d1f807c1cd405"} Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.510369 5117 generic.go:358] "Generic (PLEG): container finished" podID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerID="435ea4f262366c63d34ce3eb98403b88d5cffbb4deed5a20e239a3b28a4a95a8" exitCode=0 Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.510419 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" event={"ID":"79ef4008-b3a9-4f54-93b2-c296b982c5e7","Type":"ContainerDied","Data":"435ea4f262366c63d34ce3eb98403b88d5cffbb4deed5a20e239a3b28a4a95a8"} Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.515274 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" event={"ID":"68bdcac8-762b-4e94-8494-b571a8622dd8","Type":"ContainerStarted","Data":"a0a186c01d7f010d7ee4e5be4984042fd158b523876dce62823afbc725ef4c25"} Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.523026 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl" event={"ID":"5f884179-adb4-4b62-bec6-14f40db13be5","Type":"ContainerStarted","Data":"45807791dbb5f37ba01f581ed4f0619f70c84851038525edf83842cb02b1ac7e"} Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.542794 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567544-rmr2v" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.566908 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e836407-e9f1-4125-9810-741ae3b23431-apiservice-cert\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.567088 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/4e836407-e9f1-4125-9810-741ae3b23431-openshift-service-ca\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.567226 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e836407-e9f1-4125-9810-741ae3b23431-webhook-cert\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.567313 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-m9zh5\" (UniqueName: \"kubernetes.io/projected/4e836407-e9f1-4125-9810-741ae3b23431-kube-api-access-m9zh5\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.569140 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/4e836407-e9f1-4125-9810-741ae3b23431-openshift-service-ca\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.572015 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e836407-e9f1-4125-9810-741ae3b23431-apiservice-cert\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.577979 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e836407-e9f1-4125-9810-741ae3b23431-webhook-cert\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.604426 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9zh5\" (UniqueName: \"kubernetes.io/projected/4e836407-e9f1-4125-9810-741ae3b23431-kube-api-access-m9zh5\") pod \"perses-operator-b577b4f9-4xhsq\" (UID: \"4e836407-e9f1-4125-9810-741ae3b23431\") " pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.680422 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-7d56df5895-sgkqd"] Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.689415 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-587f9c8867-6xfv5"] Mar 21 00:24:00 crc kubenswrapper[5117]: W0321 00:24:00.694959 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0025749a_4a5d_4ee1_8de3_9dfe2093085d.slice/crio-f9a712d26fde8fb53a99ef3bf6359f670f0063414a03098f6cf86275901ab405 WatchSource:0}: Error finding container f9a712d26fde8fb53a99ef3bf6359f670f0063414a03098f6cf86275901ab405: Status 404 returned error can't find the container with id f9a712d26fde8fb53a99ef3bf6359f670f0063414a03098f6cf86275901ab405 Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.751680 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.920568 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567544-rmr2v"] Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.971960 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:24:00 crc kubenswrapper[5117]: I0321 00:24:00.972498 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:24:01 crc kubenswrapper[5117]: I0321 00:24:01.038208 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:24:01 crc kubenswrapper[5117]: I0321 00:24:01.308538 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-b577b4f9-4xhsq"] Mar 21 00:24:01 crc kubenswrapper[5117]: W0321 00:24:01.341390 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e836407_e9f1_4125_9810_741ae3b23431.slice/crio-9d803cf837cacc25fcc5ec5c604cc512d0764245697de2a5629fefc671e7516e WatchSource:0}: Error finding container 9d803cf837cacc25fcc5ec5c604cc512d0764245697de2a5629fefc671e7516e: Status 404 returned error can't find the container with id 9d803cf837cacc25fcc5ec5c604cc512d0764245697de2a5629fefc671e7516e Mar 21 00:24:01 crc kubenswrapper[5117]: I0321 00:24:01.530786 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-b577b4f9-4xhsq" event={"ID":"4e836407-e9f1-4125-9810-741ae3b23431","Type":"ContainerStarted","Data":"9d803cf837cacc25fcc5ec5c604cc512d0764245697de2a5629fefc671e7516e"} Mar 21 00:24:01 crc kubenswrapper[5117]: I0321 00:24:01.533149 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567544-rmr2v" event={"ID":"be0be9be-183f-44ae-bea1-13b4c356bd8c","Type":"ContainerStarted","Data":"1bb7ca9dd5e01e981e98bd7e251520ada5b8a7a8f43ca795d1527cbe05955397"} Mar 21 00:24:01 crc kubenswrapper[5117]: I0321 00:24:01.535403 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-587f9c8867-6xfv5" event={"ID":"61a55f96-3dbb-43c7-9749-71d00e304e57","Type":"ContainerStarted","Data":"b5d464ca658b6faac45d5c8e8c5c55dc9a962f3cd354574524848311c99b035d"} Mar 21 00:24:01 crc kubenswrapper[5117]: I0321 00:24:01.541051 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" event={"ID":"0025749a-4a5d-4ee1-8de3-9dfe2093085d","Type":"ContainerStarted","Data":"f9a712d26fde8fb53a99ef3bf6359f670f0063414a03098f6cf86275901ab405"} Mar 21 00:24:01 crc kubenswrapper[5117]: I0321 00:24:01.571642 5117 generic.go:358] "Generic (PLEG): container finished" podID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerID="2475bc14234a5c376a3cc8828dec29e5b01c59c04e03a1e50447156b0fdfa0db" exitCode=0 Mar 21 00:24:01 crc kubenswrapper[5117]: I0321 00:24:01.572104 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" event={"ID":"79ef4008-b3a9-4f54-93b2-c296b982c5e7","Type":"ContainerDied","Data":"2475bc14234a5c376a3cc8828dec29e5b01c59c04e03a1e50447156b0fdfa0db"} Mar 21 00:24:01 crc kubenswrapper[5117]: I0321 00:24:01.661588 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:24:03 crc kubenswrapper[5117]: I0321 00:24:03.002813 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:24:03 crc kubenswrapper[5117]: I0321 00:24:03.084554 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:24:03 crc kubenswrapper[5117]: I0321 00:24:03.144997 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bqhkb"] Mar 21 00:24:03 crc kubenswrapper[5117]: I0321 00:24:03.631925 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567544-rmr2v" event={"ID":"be0be9be-183f-44ae-bea1-13b4c356bd8c","Type":"ContainerStarted","Data":"49362447ce32884053fa951ea528635bce3198f9644e27f94d27b0227b67d580"} Mar 21 00:24:03 crc kubenswrapper[5117]: I0321 00:24:03.655806 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567544-rmr2v" podStartSLOduration=2.432388373 podStartE2EDuration="3.655781444s" podCreationTimestamp="2026-03-21 00:24:00 +0000 UTC" firstStartedPulling="2026-03-21 00:24:00.933019469 +0000 UTC m=+754.226306631" lastFinishedPulling="2026-03-21 00:24:02.15641253 +0000 UTC m=+755.449699702" observedRunningTime="2026-03-21 00:24:03.645503856 +0000 UTC m=+756.938791028" watchObservedRunningTime="2026-03-21 00:24:03.655781444 +0000 UTC m=+756.949068616" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.266216 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.266297 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.318067 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.679761 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" event={"ID":"79ef4008-b3a9-4f54-93b2-c296b982c5e7","Type":"ContainerDied","Data":"382c248d7f48dde9999c14e22b6dcbd0a0c90f8658d1d1d7aae2188f64f1c4aa"} Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.680124 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="382c248d7f48dde9999c14e22b6dcbd0a0c90f8658d1d1d7aae2188f64f1c4aa" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.681325 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.690673 5117 generic.go:358] "Generic (PLEG): container finished" podID="be0be9be-183f-44ae-bea1-13b4c356bd8c" containerID="49362447ce32884053fa951ea528635bce3198f9644e27f94d27b0227b67d580" exitCode=0 Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.691641 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567544-rmr2v" event={"ID":"be0be9be-183f-44ae-bea1-13b4c356bd8c","Type":"ContainerDied","Data":"49362447ce32884053fa951ea528635bce3198f9644e27f94d27b0227b67d580"} Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.691880 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bqhkb" podUID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerName="registry-server" containerID="cri-o://1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3" gracePeriod=2 Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.762003 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-bundle\") pod \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.762177 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-util\") pod \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.762357 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqvlq\" (UniqueName: \"kubernetes.io/projected/79ef4008-b3a9-4f54-93b2-c296b982c5e7-kube-api-access-nqvlq\") pod \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\" (UID: \"79ef4008-b3a9-4f54-93b2-c296b982c5e7\") " Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.765451 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-bundle" (OuterVolumeSpecName: "bundle") pod "79ef4008-b3a9-4f54-93b2-c296b982c5e7" (UID: "79ef4008-b3a9-4f54-93b2-c296b982c5e7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.775161 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-util" (OuterVolumeSpecName: "util") pod "79ef4008-b3a9-4f54-93b2-c296b982c5e7" (UID: "79ef4008-b3a9-4f54-93b2-c296b982c5e7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.791412 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79ef4008-b3a9-4f54-93b2-c296b982c5e7-kube-api-access-nqvlq" (OuterVolumeSpecName: "kube-api-access-nqvlq") pod "79ef4008-b3a9-4f54-93b2-c296b982c5e7" (UID: "79ef4008-b3a9-4f54-93b2-c296b982c5e7"). InnerVolumeSpecName "kube-api-access-nqvlq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.791546 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.864105 5117 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.864152 5117 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79ef4008-b3a9-4f54-93b2-c296b982c5e7-util\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:04 crc kubenswrapper[5117]: I0321 00:24:04.864163 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nqvlq\" (UniqueName: \"kubernetes.io/projected/79ef4008-b3a9-4f54-93b2-c296b982c5e7-kube-api-access-nqvlq\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.187161 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.268718 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bncw5\" (UniqueName: \"kubernetes.io/projected/d9d70414-581a-4fa9-89c6-54362e2659d5-kube-api-access-bncw5\") pod \"d9d70414-581a-4fa9-89c6-54362e2659d5\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.268788 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-catalog-content\") pod \"d9d70414-581a-4fa9-89c6-54362e2659d5\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.268886 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-utilities\") pod \"d9d70414-581a-4fa9-89c6-54362e2659d5\" (UID: \"d9d70414-581a-4fa9-89c6-54362e2659d5\") " Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.270148 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-utilities" (OuterVolumeSpecName: "utilities") pod "d9d70414-581a-4fa9-89c6-54362e2659d5" (UID: "d9d70414-581a-4fa9-89c6-54362e2659d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.286365 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d70414-581a-4fa9-89c6-54362e2659d5-kube-api-access-bncw5" (OuterVolumeSpecName: "kube-api-access-bncw5") pod "d9d70414-581a-4fa9-89c6-54362e2659d5" (UID: "d9d70414-581a-4fa9-89c6-54362e2659d5"). InnerVolumeSpecName "kube-api-access-bncw5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.341227 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9d70414-581a-4fa9-89c6-54362e2659d5" (UID: "d9d70414-581a-4fa9-89c6-54362e2659d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.370017 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.370057 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-bncw5\" (UniqueName: \"kubernetes.io/projected/d9d70414-581a-4fa9-89c6-54362e2659d5-kube-api-access-bncw5\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.370066 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d70414-581a-4fa9-89c6-54362e2659d5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.728685 5117 generic.go:358] "Generic (PLEG): container finished" podID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerID="1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3" exitCode=0 Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.728971 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bqhkb" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.729511 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqhkb" event={"ID":"d9d70414-581a-4fa9-89c6-54362e2659d5","Type":"ContainerDied","Data":"1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3"} Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.729540 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqhkb" event={"ID":"d9d70414-581a-4fa9-89c6-54362e2659d5","Type":"ContainerDied","Data":"fb291d257f6efb5ee572defa428aa5162acaafe4bbeebb2dcdf3cc2c8b22bbf0"} Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.729561 5117 scope.go:117] "RemoveContainer" containerID="1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.730095 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st" Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.780463 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bqhkb"] Mar 21 00:24:05 crc kubenswrapper[5117]: I0321 00:24:05.783835 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bqhkb"] Mar 21 00:24:06 crc kubenswrapper[5117]: I0321 00:24:06.740641 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqp24"] Mar 21 00:24:06 crc kubenswrapper[5117]: I0321 00:24:06.750538 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zqp24" podUID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerName="registry-server" containerID="cri-o://2575d19b634e688e0a9fbe0be767be50c2975341463d0348abbc7b7b1cfeb5f4" gracePeriod=2 Mar 21 00:24:07 crc kubenswrapper[5117]: I0321 00:24:07.610620 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d70414-581a-4fa9-89c6-54362e2659d5" path="/var/lib/kubelet/pods/d9d70414-581a-4fa9-89c6-54362e2659d5/volumes" Mar 21 00:24:07 crc kubenswrapper[5117]: I0321 00:24:07.769777 5117 generic.go:358] "Generic (PLEG): container finished" podID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerID="2575d19b634e688e0a9fbe0be767be50c2975341463d0348abbc7b7b1cfeb5f4" exitCode=0 Mar 21 00:24:07 crc kubenswrapper[5117]: I0321 00:24:07.769950 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqp24" event={"ID":"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c","Type":"ContainerDied","Data":"2575d19b634e688e0a9fbe0be767be50c2975341463d0348abbc7b7b1cfeb5f4"} Mar 21 00:24:07 crc kubenswrapper[5117]: I0321 00:24:07.948957 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-857q2"] Mar 21 00:24:07 crc kubenswrapper[5117]: I0321 00:24:07.949820 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-857q2" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="registry-server" containerID="cri-o://bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4" gracePeriod=2 Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.683040 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6"] Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.683885 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerName="util" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.683908 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerName="util" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.683933 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerName="pull" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.683941 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerName="pull" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.683981 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerName="extract" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.683989 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerName="extract" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.684005 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerName="extract-content" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.684014 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerName="extract-content" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.684024 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerName="registry-server" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.684031 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerName="registry-server" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.684045 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerName="extract-utilities" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.684053 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerName="extract-utilities" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.684174 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="d9d70414-581a-4fa9-89c6-54362e2659d5" containerName="registry-server" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.684191 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="79ef4008-b3a9-4f54-93b2-c296b982c5e7" containerName="extract" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.692114 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.695170 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"kube-root-ca.crt\"" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.695337 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"openshift-service-ca.crt\"" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.696917 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager-operator\"/\"cert-manager-operator-controller-manager-dockercfg-j82h4\"" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.700195 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6"] Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.790428 5117 generic.go:358] "Generic (PLEG): container finished" podID="35d47888-2755-479f-8c93-a27faae9a116" containerID="bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4" exitCode=0 Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.790600 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-857q2" event={"ID":"35d47888-2755-479f-8c93-a27faae9a116","Type":"ContainerDied","Data":"bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4"} Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.847695 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f0cfd0f4-805a-4875-8af6-d24e23ad310a-tmp\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pjrs6\" (UID: \"f0cfd0f4-805a-4875-8af6-d24e23ad310a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.847756 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wnmm\" (UniqueName: \"kubernetes.io/projected/f0cfd0f4-805a-4875-8af6-d24e23ad310a-kube-api-access-8wnmm\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pjrs6\" (UID: \"f0cfd0f4-805a-4875-8af6-d24e23ad310a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.950000 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f0cfd0f4-805a-4875-8af6-d24e23ad310a-tmp\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pjrs6\" (UID: \"f0cfd0f4-805a-4875-8af6-d24e23ad310a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.950100 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8wnmm\" (UniqueName: \"kubernetes.io/projected/f0cfd0f4-805a-4875-8af6-d24e23ad310a-kube-api-access-8wnmm\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pjrs6\" (UID: \"f0cfd0f4-805a-4875-8af6-d24e23ad310a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.950584 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f0cfd0f4-805a-4875-8af6-d24e23ad310a-tmp\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pjrs6\" (UID: \"f0cfd0f4-805a-4875-8af6-d24e23ad310a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" Mar 21 00:24:08 crc kubenswrapper[5117]: I0321 00:24:08.977685 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wnmm\" (UniqueName: \"kubernetes.io/projected/f0cfd0f4-805a-4875-8af6-d24e23ad310a-kube-api-access-8wnmm\") pod \"cert-manager-operator-controller-manager-7c5b8bd68-pjrs6\" (UID: \"f0cfd0f4-805a-4875-8af6-d24e23ad310a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" Mar 21 00:24:09 crc kubenswrapper[5117]: I0321 00:24:09.017623 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" Mar 21 00:24:10 crc kubenswrapper[5117]: I0321 00:24:10.956146 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567544-rmr2v" Mar 21 00:24:11 crc kubenswrapper[5117]: I0321 00:24:11.078203 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44tcb\" (UniqueName: \"kubernetes.io/projected/be0be9be-183f-44ae-bea1-13b4c356bd8c-kube-api-access-44tcb\") pod \"be0be9be-183f-44ae-bea1-13b4c356bd8c\" (UID: \"be0be9be-183f-44ae-bea1-13b4c356bd8c\") " Mar 21 00:24:11 crc kubenswrapper[5117]: I0321 00:24:11.089090 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be0be9be-183f-44ae-bea1-13b4c356bd8c-kube-api-access-44tcb" (OuterVolumeSpecName: "kube-api-access-44tcb") pod "be0be9be-183f-44ae-bea1-13b4c356bd8c" (UID: "be0be9be-183f-44ae-bea1-13b4c356bd8c"). InnerVolumeSpecName "kube-api-access-44tcb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:24:11 crc kubenswrapper[5117]: I0321 00:24:11.180455 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-44tcb\" (UniqueName: \"kubernetes.io/projected/be0be9be-183f-44ae-bea1-13b4c356bd8c-kube-api-access-44tcb\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:11 crc kubenswrapper[5117]: I0321 00:24:11.829177 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567544-rmr2v" Mar 21 00:24:11 crc kubenswrapper[5117]: I0321 00:24:11.830708 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567544-rmr2v" event={"ID":"be0be9be-183f-44ae-bea1-13b4c356bd8c","Type":"ContainerDied","Data":"1bb7ca9dd5e01e981e98bd7e251520ada5b8a7a8f43ca795d1527cbe05955397"} Mar 21 00:24:11 crc kubenswrapper[5117]: I0321 00:24:11.830775 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bb7ca9dd5e01e981e98bd7e251520ada5b8a7a8f43ca795d1527cbe05955397" Mar 21 00:24:12 crc kubenswrapper[5117]: I0321 00:24:12.011491 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567538-2vtzm"] Mar 21 00:24:12 crc kubenswrapper[5117]: I0321 00:24:12.018889 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567538-2vtzm"] Mar 21 00:24:13 crc kubenswrapper[5117]: E0321 00:24:13.004989 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4 is running failed: container process not found" containerID="bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4" cmd=["grpc_health_probe","-addr=:50051"] Mar 21 00:24:13 crc kubenswrapper[5117]: E0321 00:24:13.005834 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4 is running failed: container process not found" containerID="bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4" cmd=["grpc_health_probe","-addr=:50051"] Mar 21 00:24:13 crc kubenswrapper[5117]: E0321 00:24:13.006268 5117 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4 is running failed: container process not found" containerID="bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4" cmd=["grpc_health_probe","-addr=:50051"] Mar 21 00:24:13 crc kubenswrapper[5117]: E0321 00:24:13.006326 5117 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-857q2" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="registry-server" probeResult="unknown" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.119397 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.207428 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-catalog-content\") pod \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.207510 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-utilities\") pod \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.208676 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-utilities" (OuterVolumeSpecName: "utilities") pod "2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" (UID: "2dd6b73b-a33f-4bd3-8702-afe3e1a5542c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.276622 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" (UID: "2dd6b73b-a33f-4bd3-8702-afe3e1a5542c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.309373 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdzgd\" (UniqueName: \"kubernetes.io/projected/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-kube-api-access-tdzgd\") pod \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\" (UID: \"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c\") " Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.309637 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.309655 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.326441 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-kube-api-access-tdzgd" (OuterVolumeSpecName: "kube-api-access-tdzgd") pod "2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" (UID: "2dd6b73b-a33f-4bd3-8702-afe3e1a5542c"). InnerVolumeSpecName "kube-api-access-tdzgd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.410631 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tdzgd\" (UniqueName: \"kubernetes.io/projected/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c-kube-api-access-tdzgd\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.585439 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18e279db-6c97-4c52-b2b6-8b8f1ad0937f" path="/var/lib/kubelet/pods/18e279db-6c97-4c52-b2b6-8b8f1ad0937f/volumes" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.862755 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqp24" Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.862848 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqp24" event={"ID":"2dd6b73b-a33f-4bd3-8702-afe3e1a5542c","Type":"ContainerDied","Data":"bdb87cd77dbaee610a1e19ca9d4dceb4f271cb7beb0436c4874538b2ec41a6e7"} Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.894895 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqp24"] Mar 21 00:24:13 crc kubenswrapper[5117]: I0321 00:24:13.898394 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zqp24"] Mar 21 00:24:15 crc kubenswrapper[5117]: I0321 00:24:15.586589 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" path="/var/lib/kubelet/pods/2dd6b73b-a33f-4bd3-8702-afe3e1a5542c/volumes" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.433926 5117 scope.go:117] "RemoveContainer" containerID="3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.510949 5117 scope.go:117] "RemoveContainer" containerID="7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.581418 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.606424 5117 scope.go:117] "RemoveContainer" containerID="1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3" Mar 21 00:24:16 crc kubenswrapper[5117]: E0321 00:24:16.610401 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3\": container with ID starting with 1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3 not found: ID does not exist" containerID="1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.610454 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3"} err="failed to get container status \"1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3\": rpc error: code = NotFound desc = could not find container \"1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3\": container with ID starting with 1a4591d352f94835c431b3ed85adc8ceda2b665a7f6d46f765f8199c683952b3 not found: ID does not exist" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.610487 5117 scope.go:117] "RemoveContainer" containerID="3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006" Mar 21 00:24:16 crc kubenswrapper[5117]: E0321 00:24:16.613372 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006\": container with ID starting with 3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006 not found: ID does not exist" containerID="3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.613492 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006"} err="failed to get container status \"3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006\": rpc error: code = NotFound desc = could not find container \"3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006\": container with ID starting with 3ebbdbb936a3bee774d5d74532393e8f5fe28cf96b2b6a39cda8c8ae6beca006 not found: ID does not exist" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.613523 5117 scope.go:117] "RemoveContainer" containerID="7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86" Mar 21 00:24:16 crc kubenswrapper[5117]: E0321 00:24:16.614047 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86\": container with ID starting with 7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86 not found: ID does not exist" containerID="7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.614074 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86"} err="failed to get container status \"7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86\": rpc error: code = NotFound desc = could not find container \"7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86\": container with ID starting with 7f0d0e1ec502f590e9459a48f4c7c76b24d0f3ee1e227b1b9eb528212997af86 not found: ID does not exist" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.614090 5117 scope.go:117] "RemoveContainer" containerID="2575d19b634e688e0a9fbe0be767be50c2975341463d0348abbc7b7b1cfeb5f4" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.658307 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-utilities\") pod \"35d47888-2755-479f-8c93-a27faae9a116\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.658548 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slnb2\" (UniqueName: \"kubernetes.io/projected/35d47888-2755-479f-8c93-a27faae9a116-kube-api-access-slnb2\") pod \"35d47888-2755-479f-8c93-a27faae9a116\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.658600 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-catalog-content\") pod \"35d47888-2755-479f-8c93-a27faae9a116\" (UID: \"35d47888-2755-479f-8c93-a27faae9a116\") " Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.662701 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-utilities" (OuterVolumeSpecName: "utilities") pod "35d47888-2755-479f-8c93-a27faae9a116" (UID: "35d47888-2755-479f-8c93-a27faae9a116"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.667947 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d47888-2755-479f-8c93-a27faae9a116-kube-api-access-slnb2" (OuterVolumeSpecName: "kube-api-access-slnb2") pod "35d47888-2755-479f-8c93-a27faae9a116" (UID: "35d47888-2755-479f-8c93-a27faae9a116"). InnerVolumeSpecName "kube-api-access-slnb2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.675171 5117 scope.go:117] "RemoveContainer" containerID="ba1ae419a9a827a323a81e420056b68ddac1648f94a4a35840d8a25dcdd0c515" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.744701 5117 scope.go:117] "RemoveContainer" containerID="c518f84d824d986729678ae3f1381ba60951ffaff9dfe8a05b195fe17fcaad4c" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.765305 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-slnb2\" (UniqueName: \"kubernetes.io/projected/35d47888-2755-479f-8c93-a27faae9a116-kube-api-access-slnb2\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.765335 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.794477 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6"] Mar 21 00:24:16 crc kubenswrapper[5117]: W0321 00:24:16.797412 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0cfd0f4_805a_4875_8af6_d24e23ad310a.slice/crio-be88f72e4716372f0d8df6a79ad79df0998cb0cf4dc80dacb22afc7e8da67b4c WatchSource:0}: Error finding container be88f72e4716372f0d8df6a79ad79df0998cb0cf4dc80dacb22afc7e8da67b4c: Status 404 returned error can't find the container with id be88f72e4716372f0d8df6a79ad79df0998cb0cf4dc80dacb22afc7e8da67b4c Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.825787 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35d47888-2755-479f-8c93-a27faae9a116" (UID: "35d47888-2755-479f-8c93-a27faae9a116"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.866150 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d47888-2755-479f-8c93-a27faae9a116-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.887393 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" event={"ID":"0025749a-4a5d-4ee1-8de3-9dfe2093085d","Type":"ContainerStarted","Data":"5c36b7ba9f1b9d897fa68e68aa5ccd29bf32880db4ef842a5d0428ed56c5df9f"} Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.904180 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" event={"ID":"47f630cc-5d71-4431-8dea-397bfd008097","Type":"ContainerStarted","Data":"ef4294e8375b3377c79d8eb075b16e2e807df9de16f66e00ad972dedcfcdc8c0"} Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.908167 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-857q2" event={"ID":"35d47888-2755-479f-8c93-a27faae9a116","Type":"ContainerDied","Data":"2ebef2fbc31b4f38470a53dc5b5aaa13419b6cddad777f279c5dc4e456356900"} Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.908210 5117 scope.go:117] "RemoveContainer" containerID="bce9267821e95211cc2134526022327b60175ad2412795d829c28aaa33ad60b4" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.908356 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-857q2" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.912641 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-7d56df5895-sgkqd" podStartSLOduration=2.116266195 podStartE2EDuration="17.912624561s" podCreationTimestamp="2026-03-21 00:23:59 +0000 UTC" firstStartedPulling="2026-03-21 00:24:00.699151747 +0000 UTC m=+753.992438919" lastFinishedPulling="2026-03-21 00:24:16.495510113 +0000 UTC m=+769.788797285" observedRunningTime="2026-03-21 00:24:16.912537149 +0000 UTC m=+770.205824331" watchObservedRunningTime="2026-03-21 00:24:16.912624561 +0000 UTC m=+770.205911733" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.921761 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-b577b4f9-4xhsq" event={"ID":"4e836407-e9f1-4125-9810-741ae3b23431","Type":"ContainerStarted","Data":"22d7ee711e78a4c06acfb46c63c03c6991e931c62caac138e111d92285c70c34"} Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.921904 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.932725 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-587f9c8867-6xfv5" event={"ID":"61a55f96-3dbb-43c7-9749-71d00e304e57","Type":"ContainerStarted","Data":"996ae51f676445add42f78efbe58f87f732250c6b7b5553af172b460fe1c872b"} Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.933448 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.942461 5117 patch_prober.go:28] interesting pod/observability-operator-587f9c8867-6xfv5 container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.49:8081/healthz\": dial tcp 10.217.0.49:8081: connect: connection refused" start-of-body= Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.942529 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-587f9c8867-6xfv5" podUID="61a55f96-3dbb-43c7-9749-71d00e304e57" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.49:8081/healthz\": dial tcp 10.217.0.49:8081: connect: connection refused" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.943078 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-mggcv" podStartSLOduration=1.736791341 podStartE2EDuration="17.943060532s" podCreationTimestamp="2026-03-21 00:23:59 +0000 UTC" firstStartedPulling="2026-03-21 00:24:00.313641856 +0000 UTC m=+753.606929038" lastFinishedPulling="2026-03-21 00:24:16.519911047 +0000 UTC m=+769.813198229" observedRunningTime="2026-03-21 00:24:16.941537156 +0000 UTC m=+770.234824328" watchObservedRunningTime="2026-03-21 00:24:16.943060532 +0000 UTC m=+770.236347704" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.953493 5117 scope.go:117] "RemoveContainer" containerID="510179a5ba913b4d5b4774f36f4dba7f18843b3ae7245403fd6c473ecb3370c4" Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.965831 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl" event={"ID":"5f884179-adb4-4b62-bec6-14f40db13be5","Type":"ContainerStarted","Data":"cba157c45396ecf0f0cfcfeece546ba9a2b7fc9c46b1a5e94e6524d0c49d89e7"} Mar 21 00:24:16 crc kubenswrapper[5117]: I0321 00:24:16.967330 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" event={"ID":"f0cfd0f4-805a-4875-8af6-d24e23ad310a","Type":"ContainerStarted","Data":"be88f72e4716372f0d8df6a79ad79df0998cb0cf4dc80dacb22afc7e8da67b4c"} Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.002504 5117 scope.go:117] "RemoveContainer" containerID="549525b527c9afc6737077e66c4e3224f648e34eb607170b7dfafaf3ef200d41" Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.021923 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" podStartSLOduration=1.498223523 podStartE2EDuration="18.02189466s" podCreationTimestamp="2026-03-21 00:23:59 +0000 UTC" firstStartedPulling="2026-03-21 00:24:00.043565414 +0000 UTC m=+753.336852586" lastFinishedPulling="2026-03-21 00:24:16.567236551 +0000 UTC m=+769.860523723" observedRunningTime="2026-03-21 00:24:16.980138099 +0000 UTC m=+770.273425271" watchObservedRunningTime="2026-03-21 00:24:17.02189466 +0000 UTC m=+770.315181832" Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.022058 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-b577b4f9-4xhsq" podStartSLOduration=1.856132994 podStartE2EDuration="17.022053415s" podCreationTimestamp="2026-03-21 00:24:00 +0000 UTC" firstStartedPulling="2026-03-21 00:24:01.353845172 +0000 UTC m=+754.647132334" lastFinishedPulling="2026-03-21 00:24:16.519765573 +0000 UTC m=+769.813052755" observedRunningTime="2026-03-21 00:24:17.008743884 +0000 UTC m=+770.302031076" watchObservedRunningTime="2026-03-21 00:24:17.022053415 +0000 UTC m=+770.315340577" Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.079779 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-587f9c8867-6xfv5" podStartSLOduration=2.256684087 podStartE2EDuration="18.079759779s" podCreationTimestamp="2026-03-21 00:23:59 +0000 UTC" firstStartedPulling="2026-03-21 00:24:00.720697934 +0000 UTC m=+754.013985106" lastFinishedPulling="2026-03-21 00:24:16.543773616 +0000 UTC m=+769.837060798" observedRunningTime="2026-03-21 00:24:17.06164527 +0000 UTC m=+770.354932442" watchObservedRunningTime="2026-03-21 00:24:17.079759779 +0000 UTC m=+770.373046951" Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.082719 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-857q2"] Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.095269 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-857q2"] Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.113154 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-55568fc96c-nv4sl" podStartSLOduration=2.4722375 podStartE2EDuration="19.113133361s" podCreationTimestamp="2026-03-21 00:23:58 +0000 UTC" firstStartedPulling="2026-03-21 00:23:59.855838819 +0000 UTC m=+753.149125991" lastFinishedPulling="2026-03-21 00:24:16.49673468 +0000 UTC m=+769.790021852" observedRunningTime="2026-03-21 00:24:17.108516179 +0000 UTC m=+770.401803351" watchObservedRunningTime="2026-03-21 00:24:17.113133361 +0000 UTC m=+770.406420533" Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.596740 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d47888-2755-479f-8c93-a27faae9a116" path="/var/lib/kubelet/pods/35d47888-2755-479f-8c93-a27faae9a116/volumes" Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.980043 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5866674d46-pvrmj" event={"ID":"68bdcac8-762b-4e94-8494-b571a8622dd8","Type":"ContainerStarted","Data":"5525e0aa886bee77bccd5dace40cc067a18980da6ac11533c875fa142bc4324c"} Mar 21 00:24:17 crc kubenswrapper[5117]: I0321 00:24:17.984202 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-587f9c8867-6xfv5" Mar 21 00:24:22 crc kubenswrapper[5117]: I0321 00:24:22.010737 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" event={"ID":"f0cfd0f4-805a-4875-8af6-d24e23ad310a","Type":"ContainerStarted","Data":"7dcb28e94bab7e55995bdf5124150113048d6484f7d731023e91cc6eb6559eb8"} Mar 21 00:24:22 crc kubenswrapper[5117]: I0321 00:24:22.037624 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-7c5b8bd68-pjrs6" podStartSLOduration=9.582086423 podStartE2EDuration="14.037608089s" podCreationTimestamp="2026-03-21 00:24:08 +0000 UTC" firstStartedPulling="2026-03-21 00:24:16.803055333 +0000 UTC m=+770.096342505" lastFinishedPulling="2026-03-21 00:24:21.258576999 +0000 UTC m=+774.551864171" observedRunningTime="2026-03-21 00:24:22.03183102 +0000 UTC m=+775.325118192" watchObservedRunningTime="2026-03-21 00:24:22.037608089 +0000 UTC m=+775.330895261" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.389126 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390148 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="be0be9be-183f-44ae-bea1-13b4c356bd8c" containerName="oc" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390161 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0be9be-183f-44ae-bea1-13b4c356bd8c" containerName="oc" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390173 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="extract-utilities" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390179 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="extract-utilities" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390190 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="registry-server" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390196 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="registry-server" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390206 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerName="extract-utilities" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390212 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerName="extract-utilities" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390222 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="extract-content" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390227 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="extract-content" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390235 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerName="extract-content" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390241 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerName="extract-content" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390292 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerName="registry-server" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390298 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerName="registry-server" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390410 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="be0be9be-183f-44ae-bea1-13b4c356bd8c" containerName="oc" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390424 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="35d47888-2755-479f-8c93-a27faae9a116" containerName="registry-server" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.390432 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="2dd6b73b-a33f-4bd3-8702-afe3e1a5542c" containerName="registry-server" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.395739 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.399870 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-internal-users\"" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.400371 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-default-es-transport-certs\"" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.420812 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-default-es-config\"" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.421063 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-http-certs-internal\"" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.421349 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-xpack-file-realm\"" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.421505 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-dockercfg-nsrvt\"" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.426576 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.426695 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-remote-ca\"" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.427456 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"elasticsearch-es-unicast-hosts\"" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.427597 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"elasticsearch-es-scripts\"" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475568 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475614 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475638 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475674 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475689 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475709 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475735 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475752 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475790 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475805 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475829 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475851 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/ca7eaf9c-252d-4169-91c0-33d83751a9ae-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475865 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475891 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.475908 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577536 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577580 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577618 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577639 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577662 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577685 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/ca7eaf9c-252d-4169-91c0-33d83751a9ae-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577699 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577726 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577739 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577760 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577778 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577799 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577836 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577852 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.577872 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.579525 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.579687 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.579999 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.580300 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.580734 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.580947 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.583794 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.583872 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/ca7eaf9c-252d-4169-91c0-33d83751a9ae-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.584699 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.584894 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.589192 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.589804 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.593842 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/ca7eaf9c-252d-4169-91c0-33d83751a9ae-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.595767 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.601190 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/ca7eaf9c-252d-4169-91c0-33d83751a9ae-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"ca7eaf9c-252d-4169-91c0-33d83751a9ae\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.712525 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:23 crc kubenswrapper[5117]: I0321 00:24:23.961038 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.023423 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ca7eaf9c-252d-4169-91c0-33d83751a9ae","Type":"ContainerStarted","Data":"d28921002d512929cbb6c325326030fb6e3447870e9d294d2be941aeaa8cd812"} Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.390973 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" podUID="f85237c3-327f-4a9e-8496-00c9087e5aa7" containerName="registry" containerID="cri-o://c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79" gracePeriod=30 Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.773401 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.899809 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f85237c3-327f-4a9e-8496-00c9087e5aa7-installation-pull-secrets\") pod \"f85237c3-327f-4a9e-8496-00c9087e5aa7\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.899876 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-certificates\") pod \"f85237c3-327f-4a9e-8496-00c9087e5aa7\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.899950 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-tls\") pod \"f85237c3-327f-4a9e-8496-00c9087e5aa7\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.899994 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2677t\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-kube-api-access-2677t\") pod \"f85237c3-327f-4a9e-8496-00c9087e5aa7\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.900034 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-trusted-ca\") pod \"f85237c3-327f-4a9e-8496-00c9087e5aa7\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.900152 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-bound-sa-token\") pod \"f85237c3-327f-4a9e-8496-00c9087e5aa7\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.900189 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f85237c3-327f-4a9e-8496-00c9087e5aa7-ca-trust-extracted\") pod \"f85237c3-327f-4a9e-8496-00c9087e5aa7\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.900382 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"f85237c3-327f-4a9e-8496-00c9087e5aa7\" (UID: \"f85237c3-327f-4a9e-8496-00c9087e5aa7\") " Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.901650 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "f85237c3-327f-4a9e-8496-00c9087e5aa7" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.903308 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "f85237c3-327f-4a9e-8496-00c9087e5aa7" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.910603 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-kube-api-access-2677t" (OuterVolumeSpecName: "kube-api-access-2677t") pod "f85237c3-327f-4a9e-8496-00c9087e5aa7" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7"). InnerVolumeSpecName "kube-api-access-2677t". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.911320 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "f85237c3-327f-4a9e-8496-00c9087e5aa7" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.911667 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "f85237c3-327f-4a9e-8496-00c9087e5aa7" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.914904 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (OuterVolumeSpecName: "registry-storage") pod "f85237c3-327f-4a9e-8496-00c9087e5aa7" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7"). InnerVolumeSpecName "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2". PluginName "kubernetes.io/csi", VolumeGIDValue "" Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.920757 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f85237c3-327f-4a9e-8496-00c9087e5aa7-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "f85237c3-327f-4a9e-8496-00c9087e5aa7" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:24:24 crc kubenswrapper[5117]: I0321 00:24:24.921447 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f85237c3-327f-4a9e-8496-00c9087e5aa7-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "f85237c3-327f-4a9e-8496-00c9087e5aa7" (UID: "f85237c3-327f-4a9e-8496-00c9087e5aa7"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.002426 5117 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f85237c3-327f-4a9e-8496-00c9087e5aa7-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.002462 5117 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.002474 5117 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.002484 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2677t\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-kube-api-access-2677t\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.002492 5117 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f85237c3-327f-4a9e-8496-00c9087e5aa7-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.002502 5117 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f85237c3-327f-4a9e-8496-00c9087e5aa7-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.002510 5117 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f85237c3-327f-4a9e-8496-00c9087e5aa7-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.035015 5117 generic.go:358] "Generic (PLEG): container finished" podID="f85237c3-327f-4a9e-8496-00c9087e5aa7" containerID="c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79" exitCode=0 Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.035136 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.035275 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" event={"ID":"f85237c3-327f-4a9e-8496-00c9087e5aa7","Type":"ContainerDied","Data":"c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79"} Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.035304 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-gjth4" event={"ID":"f85237c3-327f-4a9e-8496-00c9087e5aa7","Type":"ContainerDied","Data":"b9b68b63dfd72250771d9143f4fd8af7f3e22f042dbdfa6f929caeeaf4a9f3ad"} Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.035321 5117 scope.go:117] "RemoveContainer" containerID="c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.061896 5117 scope.go:117] "RemoveContainer" containerID="c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79" Mar 21 00:24:25 crc kubenswrapper[5117]: E0321 00:24:25.062509 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79\": container with ID starting with c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79 not found: ID does not exist" containerID="c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.062535 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79"} err="failed to get container status \"c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79\": rpc error: code = NotFound desc = could not find container \"c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79\": container with ID starting with c019147630899c6e0b16a61ba3c6e4c36bbcdf65c85f66c2ae577283807e6d79 not found: ID does not exist" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.071512 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-gjth4"] Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.077296 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-gjth4"] Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.588532 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85237c3-327f-4a9e-8496-00c9087e5aa7" path="/var/lib/kubelet/pods/f85237c3-327f-4a9e-8496-00c9087e5aa7/volumes" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.982059 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-597b96b99b-7wclh"] Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.982802 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f85237c3-327f-4a9e-8496-00c9087e5aa7" containerName="registry" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.982824 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85237c3-327f-4a9e-8496-00c9087e5aa7" containerName="registry" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.982946 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="f85237c3-327f-4a9e-8496-00c9087e5aa7" containerName="registry" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.987223 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.989590 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"openshift-service-ca.crt\"" Mar 21 00:24:25 crc kubenswrapper[5117]: I0321 00:24:25.996592 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"kube-root-ca.crt\"" Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.003414 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-webhook-dockercfg-jrtnz\"" Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.008003 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-597b96b99b-7wclh"] Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.117623 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51772f05-28ae-4323-8633-c4996fb52222-bound-sa-token\") pod \"cert-manager-webhook-597b96b99b-7wclh\" (UID: \"51772f05-28ae-4323-8633-c4996fb52222\") " pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.117709 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxtb7\" (UniqueName: \"kubernetes.io/projected/51772f05-28ae-4323-8633-c4996fb52222-kube-api-access-xxtb7\") pod \"cert-manager-webhook-597b96b99b-7wclh\" (UID: \"51772f05-28ae-4323-8633-c4996fb52222\") " pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.219779 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51772f05-28ae-4323-8633-c4996fb52222-bound-sa-token\") pod \"cert-manager-webhook-597b96b99b-7wclh\" (UID: \"51772f05-28ae-4323-8633-c4996fb52222\") " pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.219892 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xxtb7\" (UniqueName: \"kubernetes.io/projected/51772f05-28ae-4323-8633-c4996fb52222-kube-api-access-xxtb7\") pod \"cert-manager-webhook-597b96b99b-7wclh\" (UID: \"51772f05-28ae-4323-8633-c4996fb52222\") " pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.241097 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51772f05-28ae-4323-8633-c4996fb52222-bound-sa-token\") pod \"cert-manager-webhook-597b96b99b-7wclh\" (UID: \"51772f05-28ae-4323-8633-c4996fb52222\") " pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.241339 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxtb7\" (UniqueName: \"kubernetes.io/projected/51772f05-28ae-4323-8633-c4996fb52222-kube-api-access-xxtb7\") pod \"cert-manager-webhook-597b96b99b-7wclh\" (UID: \"51772f05-28ae-4323-8633-c4996fb52222\") " pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.301618 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:24:26 crc kubenswrapper[5117]: I0321 00:24:26.571902 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-597b96b99b-7wclh"] Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.194298 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-8966b78d4-z7sqv"] Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.212018 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-8966b78d4-z7sqv"] Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.212172 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.215563 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-sjdwp\"" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.335067 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2aa9faf9-ec74-49e6-a0de-48c5e938bb9a-bound-sa-token\") pod \"cert-manager-cainjector-8966b78d4-z7sqv\" (UID: \"2aa9faf9-ec74-49e6-a0de-48c5e938bb9a\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.335157 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8tqj\" (UniqueName: \"kubernetes.io/projected/2aa9faf9-ec74-49e6-a0de-48c5e938bb9a-kube-api-access-w8tqj\") pod \"cert-manager-cainjector-8966b78d4-z7sqv\" (UID: \"2aa9faf9-ec74-49e6-a0de-48c5e938bb9a\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.442990 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2aa9faf9-ec74-49e6-a0de-48c5e938bb9a-bound-sa-token\") pod \"cert-manager-cainjector-8966b78d4-z7sqv\" (UID: \"2aa9faf9-ec74-49e6-a0de-48c5e938bb9a\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.443379 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-w8tqj\" (UniqueName: \"kubernetes.io/projected/2aa9faf9-ec74-49e6-a0de-48c5e938bb9a-kube-api-access-w8tqj\") pod \"cert-manager-cainjector-8966b78d4-z7sqv\" (UID: \"2aa9faf9-ec74-49e6-a0de-48c5e938bb9a\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.470997 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2aa9faf9-ec74-49e6-a0de-48c5e938bb9a-bound-sa-token\") pod \"cert-manager-cainjector-8966b78d4-z7sqv\" (UID: \"2aa9faf9-ec74-49e6-a0de-48c5e938bb9a\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.471094 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8tqj\" (UniqueName: \"kubernetes.io/projected/2aa9faf9-ec74-49e6-a0de-48c5e938bb9a-kube-api-access-w8tqj\") pod \"cert-manager-cainjector-8966b78d4-z7sqv\" (UID: \"2aa9faf9-ec74-49e6-a0de-48c5e938bb9a\") " pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.543087 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-sjdwp\"" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.552212 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" Mar 21 00:24:27 crc kubenswrapper[5117]: I0321 00:24:27.988086 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-b577b4f9-4xhsq" Mar 21 00:24:28 crc kubenswrapper[5117]: I0321 00:24:28.477220 5117 scope.go:117] "RemoveContainer" containerID="798478e891c0f903b943431f5d3dc0fc1a6e3e8f33b76532864796896db5578e" Mar 21 00:24:29 crc kubenswrapper[5117]: I0321 00:24:29.669506 5117 scope.go:117] "RemoveContainer" containerID="baac535c8d9804bf9b83f22e50e9550ac413e25ff7243825ed49e4773dc6f84a" Mar 21 00:24:29 crc kubenswrapper[5117]: I0321 00:24:29.714487 5117 scope.go:117] "RemoveContainer" containerID="196017919a8d16cbf50f101f9284c81f847e451357e39bcbc68e85c3bae37bdf" Mar 21 00:24:29 crc kubenswrapper[5117]: I0321 00:24:29.767439 5117 scope.go:117] "RemoveContainer" containerID="4a59ead9340cf7d87dc5e3a23cda40a56f9ac880489d25f65efb954c6c46f094" Mar 21 00:24:30 crc kubenswrapper[5117]: I0321 00:24:30.108976 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" event={"ID":"51772f05-28ae-4323-8633-c4996fb52222","Type":"ContainerStarted","Data":"5129dd094e3dd7d40b3872f2842a3bc91dd808c7322dfc6c14755317b94547d8"} Mar 21 00:24:30 crc kubenswrapper[5117]: I0321 00:24:30.401299 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-8966b78d4-z7sqv"] Mar 21 00:24:31 crc kubenswrapper[5117]: I0321 00:24:31.117915 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" event={"ID":"2aa9faf9-ec74-49e6-a0de-48c5e938bb9a","Type":"ContainerStarted","Data":"a96d0e106e38debc477b0416f5ca99b33b9fc9df14357880fa81570211f02a97"} Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.149075 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-759f64656b-9p6vr"] Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.280302 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-759f64656b-9p6vr"] Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.280478 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-759f64656b-9p6vr" Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.283768 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-dockercfg-bxrjk\"" Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.375972 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57fv2\" (UniqueName: \"kubernetes.io/projected/27d8650b-ad7e-452a-b1d2-5605dcfe437a-kube-api-access-57fv2\") pod \"cert-manager-759f64656b-9p6vr\" (UID: \"27d8650b-ad7e-452a-b1d2-5605dcfe437a\") " pod="cert-manager/cert-manager-759f64656b-9p6vr" Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.376423 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/27d8650b-ad7e-452a-b1d2-5605dcfe437a-bound-sa-token\") pod \"cert-manager-759f64656b-9p6vr\" (UID: \"27d8650b-ad7e-452a-b1d2-5605dcfe437a\") " pod="cert-manager/cert-manager-759f64656b-9p6vr" Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.478077 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-57fv2\" (UniqueName: \"kubernetes.io/projected/27d8650b-ad7e-452a-b1d2-5605dcfe437a-kube-api-access-57fv2\") pod \"cert-manager-759f64656b-9p6vr\" (UID: \"27d8650b-ad7e-452a-b1d2-5605dcfe437a\") " pod="cert-manager/cert-manager-759f64656b-9p6vr" Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.478690 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/27d8650b-ad7e-452a-b1d2-5605dcfe437a-bound-sa-token\") pod \"cert-manager-759f64656b-9p6vr\" (UID: \"27d8650b-ad7e-452a-b1d2-5605dcfe437a\") " pod="cert-manager/cert-manager-759f64656b-9p6vr" Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.507744 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/27d8650b-ad7e-452a-b1d2-5605dcfe437a-bound-sa-token\") pod \"cert-manager-759f64656b-9p6vr\" (UID: \"27d8650b-ad7e-452a-b1d2-5605dcfe437a\") " pod="cert-manager/cert-manager-759f64656b-9p6vr" Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.513960 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-57fv2\" (UniqueName: \"kubernetes.io/projected/27d8650b-ad7e-452a-b1d2-5605dcfe437a-kube-api-access-57fv2\") pod \"cert-manager-759f64656b-9p6vr\" (UID: \"27d8650b-ad7e-452a-b1d2-5605dcfe437a\") " pod="cert-manager/cert-manager-759f64656b-9p6vr" Mar 21 00:24:44 crc kubenswrapper[5117]: I0321 00:24:44.612499 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-759f64656b-9p6vr" Mar 21 00:24:45 crc kubenswrapper[5117]: W0321 00:24:45.465516 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27d8650b_ad7e_452a_b1d2_5605dcfe437a.slice/crio-f1e9ab3be655b8e590828b80c1605ae2df0bc3264b072b5d0f6d27d2b40b7277 WatchSource:0}: Error finding container f1e9ab3be655b8e590828b80c1605ae2df0bc3264b072b5d0f6d27d2b40b7277: Status 404 returned error can't find the container with id f1e9ab3be655b8e590828b80c1605ae2df0bc3264b072b5d0f6d27d2b40b7277 Mar 21 00:24:45 crc kubenswrapper[5117]: I0321 00:24:45.465616 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-759f64656b-9p6vr"] Mar 21 00:24:46 crc kubenswrapper[5117]: I0321 00:24:46.253241 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" event={"ID":"51772f05-28ae-4323-8633-c4996fb52222","Type":"ContainerStarted","Data":"3a00cb3b7ab4b3d8fe4c15b4e99be3bbb00a724d09493a8aa55e4da67dc141f0"} Mar 21 00:24:46 crc kubenswrapper[5117]: I0321 00:24:46.253655 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:24:46 crc kubenswrapper[5117]: I0321 00:24:46.255767 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-759f64656b-9p6vr" event={"ID":"27d8650b-ad7e-452a-b1d2-5605dcfe437a","Type":"ContainerStarted","Data":"f1e9ab3be655b8e590828b80c1605ae2df0bc3264b072b5d0f6d27d2b40b7277"} Mar 21 00:24:46 crc kubenswrapper[5117]: I0321 00:24:46.280781 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" podStartSLOduration=10.46686397 podStartE2EDuration="21.280762835s" podCreationTimestamp="2026-03-21 00:24:25 +0000 UTC" firstStartedPulling="2026-03-21 00:24:29.307473436 +0000 UTC m=+782.600760608" lastFinishedPulling="2026-03-21 00:24:40.121372301 +0000 UTC m=+793.414659473" observedRunningTime="2026-03-21 00:24:46.277712131 +0000 UTC m=+799.570999343" watchObservedRunningTime="2026-03-21 00:24:46.280762835 +0000 UTC m=+799.574050007" Mar 21 00:24:47 crc kubenswrapper[5117]: I0321 00:24:47.265554 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-759f64656b-9p6vr" event={"ID":"27d8650b-ad7e-452a-b1d2-5605dcfe437a","Type":"ContainerStarted","Data":"a27fa54c3df8221234c7b60e3e8dcf8b7fb27ded6131fc1ada9556858465d742"} Mar 21 00:24:47 crc kubenswrapper[5117]: I0321 00:24:47.268274 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" event={"ID":"2aa9faf9-ec74-49e6-a0de-48c5e938bb9a","Type":"ContainerStarted","Data":"f8c3286e1fa59341f2d2e7eecd2f72e3bd1d6c1fa12a46b22e7761db68668156"} Mar 21 00:24:47 crc kubenswrapper[5117]: I0321 00:24:47.270524 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ca7eaf9c-252d-4169-91c0-33d83751a9ae","Type":"ContainerStarted","Data":"19a17f9ef9f964d0749403c48caf22907912cbacc1c9754561df7c1c23c4f7dc"} Mar 21 00:24:47 crc kubenswrapper[5117]: I0321 00:24:47.304354 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-759f64656b-9p6vr" podStartSLOduration=3.304322261 podStartE2EDuration="3.304322261s" podCreationTimestamp="2026-03-21 00:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:24:47.290542704 +0000 UTC m=+800.583829886" watchObservedRunningTime="2026-03-21 00:24:47.304322261 +0000 UTC m=+800.597609463" Mar 21 00:24:47 crc kubenswrapper[5117]: I0321 00:24:47.330771 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-8966b78d4-z7sqv" podStartSLOduration=10.548657118 podStartE2EDuration="20.33074975s" podCreationTimestamp="2026-03-21 00:24:27 +0000 UTC" firstStartedPulling="2026-03-21 00:24:30.410839966 +0000 UTC m=+783.704127138" lastFinishedPulling="2026-03-21 00:24:40.192932588 +0000 UTC m=+793.486219770" observedRunningTime="2026-03-21 00:24:47.317768017 +0000 UTC m=+800.611055199" watchObservedRunningTime="2026-03-21 00:24:47.33074975 +0000 UTC m=+800.624036962" Mar 21 00:24:47 crc kubenswrapper[5117]: I0321 00:24:47.455436 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 21 00:24:47 crc kubenswrapper[5117]: I0321 00:24:47.517199 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 21 00:24:49 crc kubenswrapper[5117]: I0321 00:24:49.292599 5117 generic.go:358] "Generic (PLEG): container finished" podID="ca7eaf9c-252d-4169-91c0-33d83751a9ae" containerID="19a17f9ef9f964d0749403c48caf22907912cbacc1c9754561df7c1c23c4f7dc" exitCode=0 Mar 21 00:24:49 crc kubenswrapper[5117]: I0321 00:24:49.292687 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ca7eaf9c-252d-4169-91c0-33d83751a9ae","Type":"ContainerDied","Data":"19a17f9ef9f964d0749403c48caf22907912cbacc1c9754561df7c1c23c4f7dc"} Mar 21 00:24:50 crc kubenswrapper[5117]: I0321 00:24:50.308487 5117 generic.go:358] "Generic (PLEG): container finished" podID="ca7eaf9c-252d-4169-91c0-33d83751a9ae" containerID="dd8fedd87eb3480c536ad6ad95555f45a735a2f8613091a9d54ab437a76473cb" exitCode=0 Mar 21 00:24:50 crc kubenswrapper[5117]: I0321 00:24:50.308603 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ca7eaf9c-252d-4169-91c0-33d83751a9ae","Type":"ContainerDied","Data":"dd8fedd87eb3480c536ad6ad95555f45a735a2f8613091a9d54ab437a76473cb"} Mar 21 00:24:51 crc kubenswrapper[5117]: I0321 00:24:51.322161 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ca7eaf9c-252d-4169-91c0-33d83751a9ae","Type":"ContainerStarted","Data":"98ada84820836178ac43422bc64d6eb100816d680fdae56925554fadc736dd1c"} Mar 21 00:24:51 crc kubenswrapper[5117]: I0321 00:24:51.322424 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:24:51 crc kubenswrapper[5117]: I0321 00:24:51.399162 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=5.862451268 podStartE2EDuration="28.399133862s" podCreationTimestamp="2026-03-21 00:24:23 +0000 UTC" firstStartedPulling="2026-03-21 00:24:23.967865458 +0000 UTC m=+777.261152630" lastFinishedPulling="2026-03-21 00:24:46.504548052 +0000 UTC m=+799.797835224" observedRunningTime="2026-03-21 00:24:51.388295426 +0000 UTC m=+804.681582638" watchObservedRunningTime="2026-03-21 00:24:51.399133862 +0000 UTC m=+804.692421074" Mar 21 00:24:52 crc kubenswrapper[5117]: I0321 00:24:52.274499 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-597b96b99b-7wclh" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.369379 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head"] Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.383248 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head"] Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.383467 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.387502 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-catalog-configmap-partition-1\"" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.522886 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"smart-gateway-operator-catalog-configmap-partition-1-unzip\" (UniqueName: \"kubernetes.io/empty-dir/2f54262c-78bf-4ccc-a426-5f21ec87e4f6-smart-gateway-operator-catalog-configmap-partition-1-unzip\") pod \"infrawatch-operators-smart-gateway-operator-bundle-nightly-head\" (UID: \"2f54262c-78bf-4ccc-a426-5f21ec87e4f6\") " pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.522961 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"smart-gateway-operator-catalog-configmap-partition-1-volume\" (UniqueName: \"kubernetes.io/configmap/2f54262c-78bf-4ccc-a426-5f21ec87e4f6-smart-gateway-operator-catalog-configmap-partition-1-volume\") pod \"infrawatch-operators-smart-gateway-operator-bundle-nightly-head\" (UID: \"2f54262c-78bf-4ccc-a426-5f21ec87e4f6\") " pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.522998 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs2sd\" (UniqueName: \"kubernetes.io/projected/2f54262c-78bf-4ccc-a426-5f21ec87e4f6-kube-api-access-cs2sd\") pod \"infrawatch-operators-smart-gateway-operator-bundle-nightly-head\" (UID: \"2f54262c-78bf-4ccc-a426-5f21ec87e4f6\") " pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.624575 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"smart-gateway-operator-catalog-configmap-partition-1-unzip\" (UniqueName: \"kubernetes.io/empty-dir/2f54262c-78bf-4ccc-a426-5f21ec87e4f6-smart-gateway-operator-catalog-configmap-partition-1-unzip\") pod \"infrawatch-operators-smart-gateway-operator-bundle-nightly-head\" (UID: \"2f54262c-78bf-4ccc-a426-5f21ec87e4f6\") " pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.624718 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"smart-gateway-operator-catalog-configmap-partition-1-volume\" (UniqueName: \"kubernetes.io/configmap/2f54262c-78bf-4ccc-a426-5f21ec87e4f6-smart-gateway-operator-catalog-configmap-partition-1-volume\") pod \"infrawatch-operators-smart-gateway-operator-bundle-nightly-head\" (UID: \"2f54262c-78bf-4ccc-a426-5f21ec87e4f6\") " pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.624777 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cs2sd\" (UniqueName: \"kubernetes.io/projected/2f54262c-78bf-4ccc-a426-5f21ec87e4f6-kube-api-access-cs2sd\") pod \"infrawatch-operators-smart-gateway-operator-bundle-nightly-head\" (UID: \"2f54262c-78bf-4ccc-a426-5f21ec87e4f6\") " pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.625556 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"smart-gateway-operator-catalog-configmap-partition-1-unzip\" (UniqueName: \"kubernetes.io/empty-dir/2f54262c-78bf-4ccc-a426-5f21ec87e4f6-smart-gateway-operator-catalog-configmap-partition-1-unzip\") pod \"infrawatch-operators-smart-gateway-operator-bundle-nightly-head\" (UID: \"2f54262c-78bf-4ccc-a426-5f21ec87e4f6\") " pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.625942 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"smart-gateway-operator-catalog-configmap-partition-1-volume\" (UniqueName: \"kubernetes.io/configmap/2f54262c-78bf-4ccc-a426-5f21ec87e4f6-smart-gateway-operator-catalog-configmap-partition-1-volume\") pod \"infrawatch-operators-smart-gateway-operator-bundle-nightly-head\" (UID: \"2f54262c-78bf-4ccc-a426-5f21ec87e4f6\") " pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.662756 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs2sd\" (UniqueName: \"kubernetes.io/projected/2f54262c-78bf-4ccc-a426-5f21ec87e4f6-kube-api-access-cs2sd\") pod \"infrawatch-operators-smart-gateway-operator-bundle-nightly-head\" (UID: \"2f54262c-78bf-4ccc-a426-5f21ec87e4f6\") " pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:01 crc kubenswrapper[5117]: I0321 00:25:01.708886 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" Mar 21 00:25:02 crc kubenswrapper[5117]: I0321 00:25:02.310894 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head"] Mar 21 00:25:02 crc kubenswrapper[5117]: I0321 00:25:02.427555 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" event={"ID":"2f54262c-78bf-4ccc-a426-5f21ec87e4f6","Type":"ContainerStarted","Data":"ca9d1f714cdbc7b4a91ce4928c2d63ef6973d05ffe0b51f8cecb7734c154f4dd"} Mar 21 00:25:02 crc kubenswrapper[5117]: I0321 00:25:02.466075 5117 prober.go:120] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="ca7eaf9c-252d-4169-91c0-33d83751a9ae" containerName="elasticsearch" probeResult="failure" output=< Mar 21 00:25:02 crc kubenswrapper[5117]: {"timestamp": "2026-03-21T00:25:02+00:00", "message": "readiness probe failed", "curl_rc": "7"} Mar 21 00:25:02 crc kubenswrapper[5117]: > Mar 21 00:25:08 crc kubenswrapper[5117]: I0321 00:25:08.282770 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Mar 21 00:25:08 crc kubenswrapper[5117]: I0321 00:25:08.467368 5117 generic.go:358] "Generic (PLEG): container finished" podID="2f54262c-78bf-4ccc-a426-5f21ec87e4f6" containerID="513f7e04ba8b326797559abd9d1290f0a7691d851b9255c14ac1f61cd8ac193b" exitCode=0 Mar 21 00:25:08 crc kubenswrapper[5117]: I0321 00:25:08.467449 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" event={"ID":"2f54262c-78bf-4ccc-a426-5f21ec87e4f6","Type":"ContainerDied","Data":"513f7e04ba8b326797559abd9d1290f0a7691d851b9255c14ac1f61cd8ac193b"} Mar 21 00:25:11 crc kubenswrapper[5117]: I0321 00:25:11.487116 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" event={"ID":"2f54262c-78bf-4ccc-a426-5f21ec87e4f6","Type":"ContainerStarted","Data":"695371d1dcc9a01527eb36057a30b46fb739549284b2ca5b90b23b6615f64514"} Mar 21 00:25:12 crc kubenswrapper[5117]: I0321 00:25:12.517803 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-smart-gateway-operator-bundle-nightly-head" podStartSLOduration=2.5878565 podStartE2EDuration="11.517718617s" podCreationTimestamp="2026-03-21 00:25:01 +0000 UTC" firstStartedPulling="2026-03-21 00:25:02.328596768 +0000 UTC m=+815.621883930" lastFinishedPulling="2026-03-21 00:25:11.258458875 +0000 UTC m=+824.551746047" observedRunningTime="2026-03-21 00:25:12.514062353 +0000 UTC m=+825.807349555" watchObservedRunningTime="2026-03-21 00:25:12.517718617 +0000 UTC m=+825.811005829" Mar 21 00:25:14 crc kubenswrapper[5117]: I0321 00:25:14.642944 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh"] Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.167565 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh"] Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.167791 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.224389 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-bundle\") pod \"581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.224459 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vcnf\" (UniqueName: \"kubernetes.io/projected/a7170565-d2e4-4b27-aabd-d0a5784b9c79-kube-api-access-6vcnf\") pod \"581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.224724 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-util\") pod \"581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.326505 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-bundle\") pod \"581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.326567 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6vcnf\" (UniqueName: \"kubernetes.io/projected/a7170565-d2e4-4b27-aabd-d0a5784b9c79-kube-api-access-6vcnf\") pod \"581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.326627 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-util\") pod \"581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.327093 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-bundle\") pod \"581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.327136 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-util\") pod \"581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.492231 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vcnf\" (UniqueName: \"kubernetes.io/projected/a7170565-d2e4-4b27-aabd-d0a5784b9c79-kube-api-access-6vcnf\") pod \"581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.492494 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:15 crc kubenswrapper[5117]: I0321 00:25:15.723585 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh"] Mar 21 00:25:16 crc kubenswrapper[5117]: I0321 00:25:16.526775 5117 generic.go:358] "Generic (PLEG): container finished" podID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerID="aa06ca91aea36a542a148a99d9702892ba9b80268a1488fb94860fd4b6e19af2" exitCode=0 Mar 21 00:25:16 crc kubenswrapper[5117]: I0321 00:25:16.526850 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" event={"ID":"a7170565-d2e4-4b27-aabd-d0a5784b9c79","Type":"ContainerDied","Data":"aa06ca91aea36a542a148a99d9702892ba9b80268a1488fb94860fd4b6e19af2"} Mar 21 00:25:16 crc kubenswrapper[5117]: I0321 00:25:16.527234 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" event={"ID":"a7170565-d2e4-4b27-aabd-d0a5784b9c79","Type":"ContainerStarted","Data":"33f0416b653050801d9b6592b69f3fc97547d4a92b331237645daa28be0d25e5"} Mar 21 00:25:18 crc kubenswrapper[5117]: I0321 00:25:18.542390 5117 generic.go:358] "Generic (PLEG): container finished" podID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerID="ec6f5ad41a93edb3b8873fdca4fe16c4791766f612f271cd04b37024db36f34f" exitCode=0 Mar 21 00:25:18 crc kubenswrapper[5117]: I0321 00:25:18.542540 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" event={"ID":"a7170565-d2e4-4b27-aabd-d0a5784b9c79","Type":"ContainerDied","Data":"ec6f5ad41a93edb3b8873fdca4fe16c4791766f612f271cd04b37024db36f34f"} Mar 21 00:25:19 crc kubenswrapper[5117]: I0321 00:25:19.554216 5117 generic.go:358] "Generic (PLEG): container finished" podID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerID="da8bc392f98d303d6416dcf96a12ea7cad4c70cdac8b2baaed7b25f7906ac92b" exitCode=0 Mar 21 00:25:19 crc kubenswrapper[5117]: I0321 00:25:19.554340 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" event={"ID":"a7170565-d2e4-4b27-aabd-d0a5784b9c79","Type":"ContainerDied","Data":"da8bc392f98d303d6416dcf96a12ea7cad4c70cdac8b2baaed7b25f7906ac92b"} Mar 21 00:25:20 crc kubenswrapper[5117]: I0321 00:25:20.922047 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.015748 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vcnf\" (UniqueName: \"kubernetes.io/projected/a7170565-d2e4-4b27-aabd-d0a5784b9c79-kube-api-access-6vcnf\") pod \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.015875 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-util\") pod \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.016044 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-bundle\") pod \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\" (UID: \"a7170565-d2e4-4b27-aabd-d0a5784b9c79\") " Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.016916 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-bundle" (OuterVolumeSpecName: "bundle") pod "a7170565-d2e4-4b27-aabd-d0a5784b9c79" (UID: "a7170565-d2e4-4b27-aabd-d0a5784b9c79"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.022239 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7170565-d2e4-4b27-aabd-d0a5784b9c79-kube-api-access-6vcnf" (OuterVolumeSpecName: "kube-api-access-6vcnf") pod "a7170565-d2e4-4b27-aabd-d0a5784b9c79" (UID: "a7170565-d2e4-4b27-aabd-d0a5784b9c79"). InnerVolumeSpecName "kube-api-access-6vcnf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.029836 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-util" (OuterVolumeSpecName: "util") pod "a7170565-d2e4-4b27-aabd-d0a5784b9c79" (UID: "a7170565-d2e4-4b27-aabd-d0a5784b9c79"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.118233 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6vcnf\" (UniqueName: \"kubernetes.io/projected/a7170565-d2e4-4b27-aabd-d0a5784b9c79-kube-api-access-6vcnf\") on node \"crc\" DevicePath \"\"" Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.118297 5117 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-util\") on node \"crc\" DevicePath \"\"" Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.118311 5117 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7170565-d2e4-4b27-aabd-d0a5784b9c79-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.571123 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" event={"ID":"a7170565-d2e4-4b27-aabd-d0a5784b9c79","Type":"ContainerDied","Data":"33f0416b653050801d9b6592b69f3fc97547d4a92b331237645daa28be0d25e5"} Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.571626 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33f0416b653050801d9b6592b69f3fc97547d4a92b331237645daa28be0d25e5" Mar 21 00:25:21 crc kubenswrapper[5117]: I0321 00:25:21.571183 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/581064c273eeb770c9fbc3e03ee675cb542f06b12d97607b3aad976661nc6vh" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.757213 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb"] Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.758885 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerName="pull" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.758938 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerName="pull" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.759003 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerName="extract" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.759024 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerName="extract" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.759102 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerName="util" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.759126 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerName="util" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.759438 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="a7170565-d2e4-4b27-aabd-d0a5784b9c79" containerName="extract" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.771849 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.776779 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-dockercfg-fstkj\"" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.777754 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb"] Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.815733 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kz6w\" (UniqueName: \"kubernetes.io/projected/51dea3cf-4c01-4c26-99e7-d5bebf6589d4-kube-api-access-8kz6w\") pod \"smart-gateway-operator-79dd94fcb7-tkcvb\" (UID: \"51dea3cf-4c01-4c26-99e7-d5bebf6589d4\") " pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.815803 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/51dea3cf-4c01-4c26-99e7-d5bebf6589d4-runner\") pod \"smart-gateway-operator-79dd94fcb7-tkcvb\" (UID: \"51dea3cf-4c01-4c26-99e7-d5bebf6589d4\") " pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.917207 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8kz6w\" (UniqueName: \"kubernetes.io/projected/51dea3cf-4c01-4c26-99e7-d5bebf6589d4-kube-api-access-8kz6w\") pod \"smart-gateway-operator-79dd94fcb7-tkcvb\" (UID: \"51dea3cf-4c01-4c26-99e7-d5bebf6589d4\") " pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.917445 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/51dea3cf-4c01-4c26-99e7-d5bebf6589d4-runner\") pod \"smart-gateway-operator-79dd94fcb7-tkcvb\" (UID: \"51dea3cf-4c01-4c26-99e7-d5bebf6589d4\") " pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.917983 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/51dea3cf-4c01-4c26-99e7-d5bebf6589d4-runner\") pod \"smart-gateway-operator-79dd94fcb7-tkcvb\" (UID: \"51dea3cf-4c01-4c26-99e7-d5bebf6589d4\") " pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" Mar 21 00:25:24 crc kubenswrapper[5117]: I0321 00:25:24.937522 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kz6w\" (UniqueName: \"kubernetes.io/projected/51dea3cf-4c01-4c26-99e7-d5bebf6589d4-kube-api-access-8kz6w\") pod \"smart-gateway-operator-79dd94fcb7-tkcvb\" (UID: \"51dea3cf-4c01-4c26-99e7-d5bebf6589d4\") " pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" Mar 21 00:25:25 crc kubenswrapper[5117]: I0321 00:25:25.119816 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" Mar 21 00:25:25 crc kubenswrapper[5117]: I0321 00:25:25.432034 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb"] Mar 21 00:25:25 crc kubenswrapper[5117]: W0321 00:25:25.437714 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51dea3cf_4c01_4c26_99e7_d5bebf6589d4.slice/crio-3e5a273b9d32eb4b21a13e35922873cba241f23167de49bb01587b02220ab14d WatchSource:0}: Error finding container 3e5a273b9d32eb4b21a13e35922873cba241f23167de49bb01587b02220ab14d: Status 404 returned error can't find the container with id 3e5a273b9d32eb4b21a13e35922873cba241f23167de49bb01587b02220ab14d Mar 21 00:25:25 crc kubenswrapper[5117]: I0321 00:25:25.606811 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" event={"ID":"51dea3cf-4c01-4c26-99e7-d5bebf6589d4","Type":"ContainerStarted","Data":"3e5a273b9d32eb4b21a13e35922873cba241f23167de49bb01587b02220ab14d"} Mar 21 00:25:41 crc kubenswrapper[5117]: I0321 00:25:41.740556 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" event={"ID":"51dea3cf-4c01-4c26-99e7-d5bebf6589d4","Type":"ContainerStarted","Data":"75346163d6b530855d9785901f232a64a3eff50f35386b1ea683c270849e6165"} Mar 21 00:25:41 crc kubenswrapper[5117]: I0321 00:25:41.772825 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-79dd94fcb7-tkcvb" podStartSLOduration=2.030810453 podStartE2EDuration="17.772800286s" podCreationTimestamp="2026-03-21 00:25:24 +0000 UTC" firstStartedPulling="2026-03-21 00:25:25.440468385 +0000 UTC m=+838.733755557" lastFinishedPulling="2026-03-21 00:25:41.182458218 +0000 UTC m=+854.475745390" observedRunningTime="2026-03-21 00:25:41.772499457 +0000 UTC m=+855.065786649" watchObservedRunningTime="2026-03-21 00:25:41.772800286 +0000 UTC m=+855.066087468" Mar 21 00:25:54 crc kubenswrapper[5117]: I0321 00:25:54.718622 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:25:54 crc kubenswrapper[5117]: I0321 00:25:54.719679 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.615999 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head"] Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.654517 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head"] Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.654688 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.657431 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-catalog-configmap-partition-1\"" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.731066 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-operator-catalog-configmap-partition-1-unzip\" (UniqueName: \"kubernetes.io/empty-dir/2e4733d4-cede-44ae-be8b-b32f6df14029-service-telemetry-operator-catalog-configmap-partition-1-unzip\") pod \"awatch-operators-service-telemetry-operator-bundle-nightly-head\" (UID: \"2e4733d4-cede-44ae-be8b-b32f6df14029\") " pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.731560 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vg69\" (UniqueName: \"kubernetes.io/projected/2e4733d4-cede-44ae-be8b-b32f6df14029-kube-api-access-5vg69\") pod \"awatch-operators-service-telemetry-operator-bundle-nightly-head\" (UID: \"2e4733d4-cede-44ae-be8b-b32f6df14029\") " pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.731754 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-operator-catalog-configmap-partition-1-volume\" (UniqueName: \"kubernetes.io/configmap/2e4733d4-cede-44ae-be8b-b32f6df14029-service-telemetry-operator-catalog-configmap-partition-1-volume\") pod \"awatch-operators-service-telemetry-operator-bundle-nightly-head\" (UID: \"2e4733d4-cede-44ae-be8b-b32f6df14029\") " pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.833734 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-telemetry-operator-catalog-configmap-partition-1-unzip\" (UniqueName: \"kubernetes.io/empty-dir/2e4733d4-cede-44ae-be8b-b32f6df14029-service-telemetry-operator-catalog-configmap-partition-1-unzip\") pod \"awatch-operators-service-telemetry-operator-bundle-nightly-head\" (UID: \"2e4733d4-cede-44ae-be8b-b32f6df14029\") " pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.833856 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5vg69\" (UniqueName: \"kubernetes.io/projected/2e4733d4-cede-44ae-be8b-b32f6df14029-kube-api-access-5vg69\") pod \"awatch-operators-service-telemetry-operator-bundle-nightly-head\" (UID: \"2e4733d4-cede-44ae-be8b-b32f6df14029\") " pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.834012 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-telemetry-operator-catalog-configmap-partition-1-volume\" (UniqueName: \"kubernetes.io/configmap/2e4733d4-cede-44ae-be8b-b32f6df14029-service-telemetry-operator-catalog-configmap-partition-1-volume\") pod \"awatch-operators-service-telemetry-operator-bundle-nightly-head\" (UID: \"2e4733d4-cede-44ae-be8b-b32f6df14029\") " pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.834818 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-telemetry-operator-catalog-configmap-partition-1-unzip\" (UniqueName: \"kubernetes.io/empty-dir/2e4733d4-cede-44ae-be8b-b32f6df14029-service-telemetry-operator-catalog-configmap-partition-1-unzip\") pod \"awatch-operators-service-telemetry-operator-bundle-nightly-head\" (UID: \"2e4733d4-cede-44ae-be8b-b32f6df14029\") " pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.835812 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-telemetry-operator-catalog-configmap-partition-1-volume\" (UniqueName: \"kubernetes.io/configmap/2e4733d4-cede-44ae-be8b-b32f6df14029-service-telemetry-operator-catalog-configmap-partition-1-volume\") pod \"awatch-operators-service-telemetry-operator-bundle-nightly-head\" (UID: \"2e4733d4-cede-44ae-be8b-b32f6df14029\") " pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.864455 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vg69\" (UniqueName: \"kubernetes.io/projected/2e4733d4-cede-44ae-be8b-b32f6df14029-kube-api-access-5vg69\") pod \"awatch-operators-service-telemetry-operator-bundle-nightly-head\" (UID: \"2e4733d4-cede-44ae-be8b-b32f6df14029\") " pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:56 crc kubenswrapper[5117]: I0321 00:25:56.980539 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" Mar 21 00:25:57 crc kubenswrapper[5117]: I0321 00:25:57.227770 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head"] Mar 21 00:25:57 crc kubenswrapper[5117]: W0321 00:25:57.234682 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e4733d4_cede_44ae_be8b_b32f6df14029.slice/crio-597c3314cc03e3c4b61c4affb56052048b626dfb53084e47eb8a097a1a85aaa2 WatchSource:0}: Error finding container 597c3314cc03e3c4b61c4affb56052048b626dfb53084e47eb8a097a1a85aaa2: Status 404 returned error can't find the container with id 597c3314cc03e3c4b61c4affb56052048b626dfb53084e47eb8a097a1a85aaa2 Mar 21 00:25:57 crc kubenswrapper[5117]: I0321 00:25:57.898004 5117 generic.go:358] "Generic (PLEG): container finished" podID="2e4733d4-cede-44ae-be8b-b32f6df14029" containerID="3fcd4e2eb5def853fe1af51d236103c686125802e76460e40aa122b086724d25" exitCode=0 Mar 21 00:25:57 crc kubenswrapper[5117]: I0321 00:25:57.898079 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" event={"ID":"2e4733d4-cede-44ae-be8b-b32f6df14029","Type":"ContainerDied","Data":"3fcd4e2eb5def853fe1af51d236103c686125802e76460e40aa122b086724d25"} Mar 21 00:25:57 crc kubenswrapper[5117]: I0321 00:25:57.898571 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" event={"ID":"2e4733d4-cede-44ae-be8b-b32f6df14029","Type":"ContainerStarted","Data":"597c3314cc03e3c4b61c4affb56052048b626dfb53084e47eb8a097a1a85aaa2"} Mar 21 00:25:58 crc kubenswrapper[5117]: I0321 00:25:58.907446 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" event={"ID":"2e4733d4-cede-44ae-be8b-b32f6df14029","Type":"ContainerStarted","Data":"83064a68e63884391e54d25f169e9ff607ddd8ae20f7d6eb4b7f60051a1e53f9"} Mar 21 00:25:58 crc kubenswrapper[5117]: I0321 00:25:58.934634 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/awatch-operators-service-telemetry-operator-bundle-nightly-head" podStartSLOduration=2.435795996 podStartE2EDuration="2.934615918s" podCreationTimestamp="2026-03-21 00:25:56 +0000 UTC" firstStartedPulling="2026-03-21 00:25:57.902922369 +0000 UTC m=+871.196209581" lastFinishedPulling="2026-03-21 00:25:58.401742291 +0000 UTC m=+871.695029503" observedRunningTime="2026-03-21 00:25:58.933643557 +0000 UTC m=+872.226930739" watchObservedRunningTime="2026-03-21 00:25:58.934615918 +0000 UTC m=+872.227903090" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.868474 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v"] Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.885041 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v"] Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.885457 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.889524 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-b2ccr\"" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.896062 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.896128 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv9f8\" (UniqueName: \"kubernetes.io/projected/364d7a25-35e5-49c8-b04a-79706601c454-kube-api-access-xv9f8\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.896451 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.998035 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.998181 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.998219 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xv9f8\" (UniqueName: \"kubernetes.io/projected/364d7a25-35e5-49c8-b04a-79706601c454-kube-api-access-xv9f8\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.998634 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:25:59 crc kubenswrapper[5117]: I0321 00:25:59.998938 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.026046 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv9f8\" (UniqueName: \"kubernetes.io/projected/364d7a25-35e5-49c8-b04a-79706601c454-kube-api-access-xv9f8\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.154413 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567546-f7pnm"] Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.163710 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567546-f7pnm" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.165123 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567546-f7pnm"] Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.171130 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.171643 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.171857 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.201162 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s7tf\" (UniqueName: \"kubernetes.io/projected/76525abe-9640-42c1-9871-bab8a511ddf0-kube-api-access-6s7tf\") pod \"auto-csr-approver-29567546-f7pnm\" (UID: \"76525abe-9640-42c1-9871-bab8a511ddf0\") " pod="openshift-infra/auto-csr-approver-29567546-f7pnm" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.214715 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.302432 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6s7tf\" (UniqueName: \"kubernetes.io/projected/76525abe-9640-42c1-9871-bab8a511ddf0-kube-api-access-6s7tf\") pod \"auto-csr-approver-29567546-f7pnm\" (UID: \"76525abe-9640-42c1-9871-bab8a511ddf0\") " pod="openshift-infra/auto-csr-approver-29567546-f7pnm" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.331957 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s7tf\" (UniqueName: \"kubernetes.io/projected/76525abe-9640-42c1-9871-bab8a511ddf0-kube-api-access-6s7tf\") pod \"auto-csr-approver-29567546-f7pnm\" (UID: \"76525abe-9640-42c1-9871-bab8a511ddf0\") " pod="openshift-infra/auto-csr-approver-29567546-f7pnm" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.472216 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v"] Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.485826 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567546-f7pnm" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.641876 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7"] Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.656648 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7"] Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.656801 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.708141 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-util\") pod \"59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.708205 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-bundle\") pod \"59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.708315 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt47d\" (UniqueName: \"kubernetes.io/projected/9a583740-0e88-4d62-bfcb-2106deac4761-kube-api-access-nt47d\") pod \"59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: W0321 00:26:00.742028 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76525abe_9640_42c1_9871_bab8a511ddf0.slice/crio-88d22f8d9573dee9398aa197cd7c00a562165d6161660150260360826a1a080e WatchSource:0}: Error finding container 88d22f8d9573dee9398aa197cd7c00a562165d6161660150260360826a1a080e: Status 404 returned error can't find the container with id 88d22f8d9573dee9398aa197cd7c00a562165d6161660150260360826a1a080e Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.744096 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567546-f7pnm"] Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.809360 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-util\") pod \"59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.809451 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-bundle\") pod \"59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.809578 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nt47d\" (UniqueName: \"kubernetes.io/projected/9a583740-0e88-4d62-bfcb-2106deac4761-kube-api-access-nt47d\") pod \"59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.810607 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-bundle\") pod \"59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.810689 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-util\") pod \"59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.838452 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt47d\" (UniqueName: \"kubernetes.io/projected/9a583740-0e88-4d62-bfcb-2106deac4761-kube-api-access-nt47d\") pod \"59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.923677 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567546-f7pnm" event={"ID":"76525abe-9640-42c1-9871-bab8a511ddf0","Type":"ContainerStarted","Data":"88d22f8d9573dee9398aa197cd7c00a562165d6161660150260360826a1a080e"} Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.925737 5117 generic.go:358] "Generic (PLEG): container finished" podID="364d7a25-35e5-49c8-b04a-79706601c454" containerID="750862935b07b863f4d10cbf814e11bc091f270d96378ab656c1eac58f288897" exitCode=0 Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.925779 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" event={"ID":"364d7a25-35e5-49c8-b04a-79706601c454","Type":"ContainerDied","Data":"750862935b07b863f4d10cbf814e11bc091f270d96378ab656c1eac58f288897"} Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.925822 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" event={"ID":"364d7a25-35e5-49c8-b04a-79706601c454","Type":"ContainerStarted","Data":"cc7074824b2b9ea8e7037ee1d347db5cbfca7c8f5a69eca45a13852452dabba8"} Mar 21 00:26:00 crc kubenswrapper[5117]: I0321 00:26:00.984509 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:01 crc kubenswrapper[5117]: I0321 00:26:01.215235 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7"] Mar 21 00:26:01 crc kubenswrapper[5117]: W0321 00:26:01.226302 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a583740_0e88_4d62_bfcb_2106deac4761.slice/crio-f7d5e3b7939e79b06db8ebcb056e8a1e2fa1c698fd18c5a384edca1b6c87493d WatchSource:0}: Error finding container f7d5e3b7939e79b06db8ebcb056e8a1e2fa1c698fd18c5a384edca1b6c87493d: Status 404 returned error can't find the container with id f7d5e3b7939e79b06db8ebcb056e8a1e2fa1c698fd18c5a384edca1b6c87493d Mar 21 00:26:01 crc kubenswrapper[5117]: I0321 00:26:01.946447 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567546-f7pnm" event={"ID":"76525abe-9640-42c1-9871-bab8a511ddf0","Type":"ContainerStarted","Data":"2a14a9daae14497a58fa06ab75253f0f3105d56bd9250653366c77b511c09b3e"} Mar 21 00:26:01 crc kubenswrapper[5117]: I0321 00:26:01.949374 5117 generic.go:358] "Generic (PLEG): container finished" podID="9a583740-0e88-4d62-bfcb-2106deac4761" containerID="00a22c1782b455cf553bcf95d24ee1eab0490aabb8c059a922d7826c03111a9c" exitCode=0 Mar 21 00:26:01 crc kubenswrapper[5117]: I0321 00:26:01.949428 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" event={"ID":"9a583740-0e88-4d62-bfcb-2106deac4761","Type":"ContainerDied","Data":"00a22c1782b455cf553bcf95d24ee1eab0490aabb8c059a922d7826c03111a9c"} Mar 21 00:26:01 crc kubenswrapper[5117]: I0321 00:26:01.949496 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" event={"ID":"9a583740-0e88-4d62-bfcb-2106deac4761","Type":"ContainerStarted","Data":"f7d5e3b7939e79b06db8ebcb056e8a1e2fa1c698fd18c5a384edca1b6c87493d"} Mar 21 00:26:01 crc kubenswrapper[5117]: I0321 00:26:01.965162 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567546-f7pnm" podStartSLOduration=1.227335341 podStartE2EDuration="1.96514025s" podCreationTimestamp="2026-03-21 00:26:00 +0000 UTC" firstStartedPulling="2026-03-21 00:26:00.743056321 +0000 UTC m=+874.036343503" lastFinishedPulling="2026-03-21 00:26:01.48086124 +0000 UTC m=+874.774148412" observedRunningTime="2026-03-21 00:26:01.964330036 +0000 UTC m=+875.257617228" watchObservedRunningTime="2026-03-21 00:26:01.96514025 +0000 UTC m=+875.258427462" Mar 21 00:26:02 crc kubenswrapper[5117]: I0321 00:26:02.960056 5117 generic.go:358] "Generic (PLEG): container finished" podID="76525abe-9640-42c1-9871-bab8a511ddf0" containerID="2a14a9daae14497a58fa06ab75253f0f3105d56bd9250653366c77b511c09b3e" exitCode=0 Mar 21 00:26:02 crc kubenswrapper[5117]: I0321 00:26:02.960479 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567546-f7pnm" event={"ID":"76525abe-9640-42c1-9871-bab8a511ddf0","Type":"ContainerDied","Data":"2a14a9daae14497a58fa06ab75253f0f3105d56bd9250653366c77b511c09b3e"} Mar 21 00:26:02 crc kubenswrapper[5117]: I0321 00:26:02.964244 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" event={"ID":"9a583740-0e88-4d62-bfcb-2106deac4761","Type":"ContainerStarted","Data":"55b79ee2c690b556ebc8838d8e709cb0dc41d01cafe09188c27e2a7d1c0938db"} Mar 21 00:26:02 crc kubenswrapper[5117]: I0321 00:26:02.968689 5117 generic.go:358] "Generic (PLEG): container finished" podID="364d7a25-35e5-49c8-b04a-79706601c454" containerID="0a56810638ea51f7ccea168444c05b2b020ad3bff22c26408eed416fa6fb83b2" exitCode=0 Mar 21 00:26:02 crc kubenswrapper[5117]: I0321 00:26:02.968807 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" event={"ID":"364d7a25-35e5-49c8-b04a-79706601c454","Type":"ContainerDied","Data":"0a56810638ea51f7ccea168444c05b2b020ad3bff22c26408eed416fa6fb83b2"} Mar 21 00:26:03 crc kubenswrapper[5117]: I0321 00:26:03.976597 5117 generic.go:358] "Generic (PLEG): container finished" podID="364d7a25-35e5-49c8-b04a-79706601c454" containerID="2bcbff705d251e5dd56a90d8687d550e57973e9c9def60341abb29da17d4e98c" exitCode=0 Mar 21 00:26:03 crc kubenswrapper[5117]: I0321 00:26:03.976697 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" event={"ID":"364d7a25-35e5-49c8-b04a-79706601c454","Type":"ContainerDied","Data":"2bcbff705d251e5dd56a90d8687d550e57973e9c9def60341abb29da17d4e98c"} Mar 21 00:26:03 crc kubenswrapper[5117]: I0321 00:26:03.979411 5117 generic.go:358] "Generic (PLEG): container finished" podID="9a583740-0e88-4d62-bfcb-2106deac4761" containerID="55b79ee2c690b556ebc8838d8e709cb0dc41d01cafe09188c27e2a7d1c0938db" exitCode=0 Mar 21 00:26:03 crc kubenswrapper[5117]: I0321 00:26:03.979443 5117 generic.go:358] "Generic (PLEG): container finished" podID="9a583740-0e88-4d62-bfcb-2106deac4761" containerID="28529e9c7be76063b6b3ba1868ae28c04f4a9f4e5fbe40ef882420313104865e" exitCode=0 Mar 21 00:26:03 crc kubenswrapper[5117]: I0321 00:26:03.979559 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" event={"ID":"9a583740-0e88-4d62-bfcb-2106deac4761","Type":"ContainerDied","Data":"55b79ee2c690b556ebc8838d8e709cb0dc41d01cafe09188c27e2a7d1c0938db"} Mar 21 00:26:03 crc kubenswrapper[5117]: I0321 00:26:03.979628 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" event={"ID":"9a583740-0e88-4d62-bfcb-2106deac4761","Type":"ContainerDied","Data":"28529e9c7be76063b6b3ba1868ae28c04f4a9f4e5fbe40ef882420313104865e"} Mar 21 00:26:04 crc kubenswrapper[5117]: I0321 00:26:04.261268 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567546-f7pnm" Mar 21 00:26:04 crc kubenswrapper[5117]: I0321 00:26:04.364543 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s7tf\" (UniqueName: \"kubernetes.io/projected/76525abe-9640-42c1-9871-bab8a511ddf0-kube-api-access-6s7tf\") pod \"76525abe-9640-42c1-9871-bab8a511ddf0\" (UID: \"76525abe-9640-42c1-9871-bab8a511ddf0\") " Mar 21 00:26:04 crc kubenswrapper[5117]: I0321 00:26:04.371808 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76525abe-9640-42c1-9871-bab8a511ddf0-kube-api-access-6s7tf" (OuterVolumeSpecName: "kube-api-access-6s7tf") pod "76525abe-9640-42c1-9871-bab8a511ddf0" (UID: "76525abe-9640-42c1-9871-bab8a511ddf0"). InnerVolumeSpecName "kube-api-access-6s7tf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:26:04 crc kubenswrapper[5117]: I0321 00:26:04.466215 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6s7tf\" (UniqueName: \"kubernetes.io/projected/76525abe-9640-42c1-9871-bab8a511ddf0-kube-api-access-6s7tf\") on node \"crc\" DevicePath \"\"" Mar 21 00:26:04 crc kubenswrapper[5117]: I0321 00:26:04.991005 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567546-f7pnm" Mar 21 00:26:04 crc kubenswrapper[5117]: I0321 00:26:04.991103 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567546-f7pnm" event={"ID":"76525abe-9640-42c1-9871-bab8a511ddf0","Type":"ContainerDied","Data":"88d22f8d9573dee9398aa197cd7c00a562165d6161660150260360826a1a080e"} Mar 21 00:26:04 crc kubenswrapper[5117]: I0321 00:26:04.991150 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88d22f8d9573dee9398aa197cd7c00a562165d6161660150260360826a1a080e" Mar 21 00:26:05 crc kubenswrapper[5117]: I0321 00:26:05.328567 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567540-4pg74"] Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.332500 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567540-4pg74"] Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.556441 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.562607 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.591439 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c9e3f55-a1b5-45fd-989c-b9c919502c57" path="/var/lib/kubelet/pods/7c9e3f55-a1b5-45fd-989c-b9c919502c57/volumes" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.685277 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-bundle\") pod \"9a583740-0e88-4d62-bfcb-2106deac4761\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.685382 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-util\") pod \"9a583740-0e88-4d62-bfcb-2106deac4761\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.685445 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-bundle\") pod \"364d7a25-35e5-49c8-b04a-79706601c454\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.685603 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt47d\" (UniqueName: \"kubernetes.io/projected/9a583740-0e88-4d62-bfcb-2106deac4761-kube-api-access-nt47d\") pod \"9a583740-0e88-4d62-bfcb-2106deac4761\" (UID: \"9a583740-0e88-4d62-bfcb-2106deac4761\") " Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.685831 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv9f8\" (UniqueName: \"kubernetes.io/projected/364d7a25-35e5-49c8-b04a-79706601c454-kube-api-access-xv9f8\") pod \"364d7a25-35e5-49c8-b04a-79706601c454\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.685935 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-util\") pod \"364d7a25-35e5-49c8-b04a-79706601c454\" (UID: \"364d7a25-35e5-49c8-b04a-79706601c454\") " Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.688655 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-bundle" (OuterVolumeSpecName: "bundle") pod "9a583740-0e88-4d62-bfcb-2106deac4761" (UID: "9a583740-0e88-4d62-bfcb-2106deac4761"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.692869 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-bundle" (OuterVolumeSpecName: "bundle") pod "364d7a25-35e5-49c8-b04a-79706601c454" (UID: "364d7a25-35e5-49c8-b04a-79706601c454"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.698452 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a583740-0e88-4d62-bfcb-2106deac4761-kube-api-access-nt47d" (OuterVolumeSpecName: "kube-api-access-nt47d") pod "9a583740-0e88-4d62-bfcb-2106deac4761" (UID: "9a583740-0e88-4d62-bfcb-2106deac4761"). InnerVolumeSpecName "kube-api-access-nt47d". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.707880 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-util" (OuterVolumeSpecName: "util") pod "9a583740-0e88-4d62-bfcb-2106deac4761" (UID: "9a583740-0e88-4d62-bfcb-2106deac4761"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.708911 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-util" (OuterVolumeSpecName: "util") pod "364d7a25-35e5-49c8-b04a-79706601c454" (UID: "364d7a25-35e5-49c8-b04a-79706601c454"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.712499 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/364d7a25-35e5-49c8-b04a-79706601c454-kube-api-access-xv9f8" (OuterVolumeSpecName: "kube-api-access-xv9f8") pod "364d7a25-35e5-49c8-b04a-79706601c454" (UID: "364d7a25-35e5-49c8-b04a-79706601c454"). InnerVolumeSpecName "kube-api-access-xv9f8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.787047 5117 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-util\") on node \"crc\" DevicePath \"\"" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.787073 5117 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.787084 5117 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a583740-0e88-4d62-bfcb-2106deac4761-util\") on node \"crc\" DevicePath \"\"" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.787154 5117 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/364d7a25-35e5-49c8-b04a-79706601c454-bundle\") on node \"crc\" DevicePath \"\"" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.787165 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nt47d\" (UniqueName: \"kubernetes.io/projected/9a583740-0e88-4d62-bfcb-2106deac4761-kube-api-access-nt47d\") on node \"crc\" DevicePath \"\"" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:05.787176 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xv9f8\" (UniqueName: \"kubernetes.io/projected/364d7a25-35e5-49c8-b04a-79706601c454-kube-api-access-xv9f8\") on node \"crc\" DevicePath \"\"" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:06.007083 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:06.007081 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/59d91eeadfbc177692af3c8c1571c9d473bd01e833d0373cf802b3d5727b4m7" event={"ID":"9a583740-0e88-4d62-bfcb-2106deac4761","Type":"ContainerDied","Data":"f7d5e3b7939e79b06db8ebcb056e8a1e2fa1c698fd18c5a384edca1b6c87493d"} Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:06.007318 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7d5e3b7939e79b06db8ebcb056e8a1e2fa1c698fd18c5a384edca1b6c87493d" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:06.010622 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:06.010666 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v" event={"ID":"364d7a25-35e5-49c8-b04a-79706601c454","Type":"ContainerDied","Data":"cc7074824b2b9ea8e7037ee1d347db5cbfca7c8f5a69eca45a13852452dabba8"} Mar 21 00:26:06 crc kubenswrapper[5117]: I0321 00:26:06.010705 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc7074824b2b9ea8e7037ee1d347db5cbfca7c8f5a69eca45a13852452dabba8" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.958020 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-b555588b5-tfz84"] Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959500 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9a583740-0e88-4d62-bfcb-2106deac4761" containerName="pull" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959527 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a583740-0e88-4d62-bfcb-2106deac4761" containerName="pull" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959550 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="76525abe-9640-42c1-9871-bab8a511ddf0" containerName="oc" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959562 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="76525abe-9640-42c1-9871-bab8a511ddf0" containerName="oc" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959592 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="364d7a25-35e5-49c8-b04a-79706601c454" containerName="pull" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959604 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="364d7a25-35e5-49c8-b04a-79706601c454" containerName="pull" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959640 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="364d7a25-35e5-49c8-b04a-79706601c454" containerName="extract" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959652 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="364d7a25-35e5-49c8-b04a-79706601c454" containerName="extract" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959667 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9a583740-0e88-4d62-bfcb-2106deac4761" containerName="util" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959677 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a583740-0e88-4d62-bfcb-2106deac4761" containerName="util" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959696 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9a583740-0e88-4d62-bfcb-2106deac4761" containerName="extract" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959707 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a583740-0e88-4d62-bfcb-2106deac4761" containerName="extract" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959734 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="364d7a25-35e5-49c8-b04a-79706601c454" containerName="util" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959744 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="364d7a25-35e5-49c8-b04a-79706601c454" containerName="util" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959904 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="364d7a25-35e5-49c8-b04a-79706601c454" containerName="extract" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959924 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="76525abe-9640-42c1-9871-bab8a511ddf0" containerName="oc" Mar 21 00:26:10 crc kubenswrapper[5117]: I0321 00:26:10.959949 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="9a583740-0e88-4d62-bfcb-2106deac4761" containerName="extract" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.282319 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-b555588b5-tfz84"] Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.282451 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.284359 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-dockercfg-ttzvr\"" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.399228 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/647ba439-f6c5-4a74-890f-28f776a83520-runner\") pod \"service-telemetry-operator-b555588b5-tfz84\" (UID: \"647ba439-f6c5-4a74-890f-28f776a83520\") " pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.399353 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cvbx\" (UniqueName: \"kubernetes.io/projected/647ba439-f6c5-4a74-890f-28f776a83520-kube-api-access-4cvbx\") pod \"service-telemetry-operator-b555588b5-tfz84\" (UID: \"647ba439-f6c5-4a74-890f-28f776a83520\") " pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.500764 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/647ba439-f6c5-4a74-890f-28f776a83520-runner\") pod \"service-telemetry-operator-b555588b5-tfz84\" (UID: \"647ba439-f6c5-4a74-890f-28f776a83520\") " pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.500840 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4cvbx\" (UniqueName: \"kubernetes.io/projected/647ba439-f6c5-4a74-890f-28f776a83520-kube-api-access-4cvbx\") pod \"service-telemetry-operator-b555588b5-tfz84\" (UID: \"647ba439-f6c5-4a74-890f-28f776a83520\") " pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.501378 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/647ba439-f6c5-4a74-890f-28f776a83520-runner\") pod \"service-telemetry-operator-b555588b5-tfz84\" (UID: \"647ba439-f6c5-4a74-890f-28f776a83520\") " pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.525644 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cvbx\" (UniqueName: \"kubernetes.io/projected/647ba439-f6c5-4a74-890f-28f776a83520-kube-api-access-4cvbx\") pod \"service-telemetry-operator-b555588b5-tfz84\" (UID: \"647ba439-f6c5-4a74-890f-28f776a83520\") " pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.598276 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" Mar 21 00:26:11 crc kubenswrapper[5117]: I0321 00:26:11.886808 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-b555588b5-tfz84"] Mar 21 00:26:11 crc kubenswrapper[5117]: W0321 00:26:11.889946 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod647ba439_f6c5_4a74_890f_28f776a83520.slice/crio-94c6a27cbf3b4d8b3f49cbaacb24677e1ee18dae7001d5fd4a0c603823434a08 WatchSource:0}: Error finding container 94c6a27cbf3b4d8b3f49cbaacb24677e1ee18dae7001d5fd4a0c603823434a08: Status 404 returned error can't find the container with id 94c6a27cbf3b4d8b3f49cbaacb24677e1ee18dae7001d5fd4a0c603823434a08 Mar 21 00:26:12 crc kubenswrapper[5117]: I0321 00:26:12.077542 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" event={"ID":"647ba439-f6c5-4a74-890f-28f776a83520","Type":"ContainerStarted","Data":"94c6a27cbf3b4d8b3f49cbaacb24677e1ee18dae7001d5fd4a0c603823434a08"} Mar 21 00:26:12 crc kubenswrapper[5117]: I0321 00:26:12.665418 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-78b9bd8798-hnjj9"] Mar 21 00:26:12 crc kubenswrapper[5117]: I0321 00:26:12.687436 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-78b9bd8798-hnjj9" Mar 21 00:26:12 crc kubenswrapper[5117]: I0321 00:26:12.692232 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"interconnect-operator-dockercfg-95lwk\"" Mar 21 00:26:12 crc kubenswrapper[5117]: I0321 00:26:12.703553 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-78b9bd8798-hnjj9"] Mar 21 00:26:12 crc kubenswrapper[5117]: I0321 00:26:12.842770 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7q2m\" (UniqueName: \"kubernetes.io/projected/a9d52748-f247-403a-a5c8-5ea458eea2b1-kube-api-access-l7q2m\") pod \"interconnect-operator-78b9bd8798-hnjj9\" (UID: \"a9d52748-f247-403a-a5c8-5ea458eea2b1\") " pod="service-telemetry/interconnect-operator-78b9bd8798-hnjj9" Mar 21 00:26:12 crc kubenswrapper[5117]: I0321 00:26:12.944927 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7q2m\" (UniqueName: \"kubernetes.io/projected/a9d52748-f247-403a-a5c8-5ea458eea2b1-kube-api-access-l7q2m\") pod \"interconnect-operator-78b9bd8798-hnjj9\" (UID: \"a9d52748-f247-403a-a5c8-5ea458eea2b1\") " pod="service-telemetry/interconnect-operator-78b9bd8798-hnjj9" Mar 21 00:26:12 crc kubenswrapper[5117]: I0321 00:26:12.975604 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7q2m\" (UniqueName: \"kubernetes.io/projected/a9d52748-f247-403a-a5c8-5ea458eea2b1-kube-api-access-l7q2m\") pod \"interconnect-operator-78b9bd8798-hnjj9\" (UID: \"a9d52748-f247-403a-a5c8-5ea458eea2b1\") " pod="service-telemetry/interconnect-operator-78b9bd8798-hnjj9" Mar 21 00:26:13 crc kubenswrapper[5117]: I0321 00:26:13.027975 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-78b9bd8798-hnjj9" Mar 21 00:26:13 crc kubenswrapper[5117]: I0321 00:26:13.261742 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-78b9bd8798-hnjj9"] Mar 21 00:26:13 crc kubenswrapper[5117]: W0321 00:26:13.269432 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9d52748_f247_403a_a5c8_5ea458eea2b1.slice/crio-fb6305801b7ace06271dd43063980a3d6c02b5f356dea1130286b99f17d148f1 WatchSource:0}: Error finding container fb6305801b7ace06271dd43063980a3d6c02b5f356dea1130286b99f17d148f1: Status 404 returned error can't find the container with id fb6305801b7ace06271dd43063980a3d6c02b5f356dea1130286b99f17d148f1 Mar 21 00:26:14 crc kubenswrapper[5117]: I0321 00:26:14.093821 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-78b9bd8798-hnjj9" event={"ID":"a9d52748-f247-403a-a5c8-5ea458eea2b1","Type":"ContainerStarted","Data":"fb6305801b7ace06271dd43063980a3d6c02b5f356dea1130286b99f17d148f1"} Mar 21 00:26:23 crc kubenswrapper[5117]: I0321 00:26:23.174316 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-78b9bd8798-hnjj9" event={"ID":"a9d52748-f247-403a-a5c8-5ea458eea2b1","Type":"ContainerStarted","Data":"9e96eb4053599a3c6c4e0e3dc3abee74ec081f9fe3a4b7682728214c5eec2387"} Mar 21 00:26:23 crc kubenswrapper[5117]: I0321 00:26:23.175796 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" event={"ID":"647ba439-f6c5-4a74-890f-28f776a83520","Type":"ContainerStarted","Data":"9c3d656b81408e8631ad9cd584055021377ac7187fe34e15e995c12fd3957baf"} Mar 21 00:26:23 crc kubenswrapper[5117]: I0321 00:26:23.196452 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-78b9bd8798-hnjj9" podStartSLOduration=1.979329929 podStartE2EDuration="11.196434809s" podCreationTimestamp="2026-03-21 00:26:12 +0000 UTC" firstStartedPulling="2026-03-21 00:26:13.272238561 +0000 UTC m=+886.565525733" lastFinishedPulling="2026-03-21 00:26:22.489343431 +0000 UTC m=+895.782630613" observedRunningTime="2026-03-21 00:26:23.189978308 +0000 UTC m=+896.483265490" watchObservedRunningTime="2026-03-21 00:26:23.196434809 +0000 UTC m=+896.489722001" Mar 21 00:26:23 crc kubenswrapper[5117]: I0321 00:26:23.219270 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-b555588b5-tfz84" podStartSLOduration=2.592534123 podStartE2EDuration="13.219241385s" podCreationTimestamp="2026-03-21 00:26:10 +0000 UTC" firstStartedPulling="2026-03-21 00:26:11.891576607 +0000 UTC m=+885.184863779" lastFinishedPulling="2026-03-21 00:26:22.518283869 +0000 UTC m=+895.811571041" observedRunningTime="2026-03-21 00:26:23.209233705 +0000 UTC m=+896.502520917" watchObservedRunningTime="2026-03-21 00:26:23.219241385 +0000 UTC m=+896.512528547" Mar 21 00:26:24 crc kubenswrapper[5117]: I0321 00:26:24.719339 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:26:24 crc kubenswrapper[5117]: I0321 00:26:24.720811 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:26:28 crc kubenswrapper[5117]: I0321 00:26:28.064035 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-72hnj_c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e/kube-multus/0.log" Mar 21 00:26:28 crc kubenswrapper[5117]: I0321 00:26:28.070807 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-72hnj_c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e/kube-multus/0.log" Mar 21 00:26:28 crc kubenswrapper[5117]: I0321 00:26:28.076130 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:26:28 crc kubenswrapper[5117]: I0321 00:26:28.079605 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:26:29 crc kubenswrapper[5117]: I0321 00:26:29.978580 5117 scope.go:117] "RemoveContainer" containerID="e5763f91a5427b8cdd0cdc1e2e284f9c3cb0ca638a2ab7cc5c6c092f8749de22" Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.882042 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-rlct4"] Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.897943 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.901976 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-users\"" Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.902026 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-interconnect-sasl-config\"" Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.902500 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-openstack-credentials\"" Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.903012 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-inter-router-ca\"" Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.903276 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-openstack-ca\"" Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.903478 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-inter-router-credentials\"" Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.903712 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-dockercfg-4j5qq\"" Mar 21 00:26:43 crc kubenswrapper[5117]: I0321 00:26:43.911409 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-rlct4"] Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.023081 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.023175 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-users\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.023203 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.023221 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jghc6\" (UniqueName: \"kubernetes.io/projected/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-kube-api-access-jghc6\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.023239 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.023274 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-config\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.023328 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.124810 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.124867 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.124902 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-users\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.124933 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.124951 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jghc6\" (UniqueName: \"kubernetes.io/projected/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-kube-api-access-jghc6\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.124969 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.124990 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-config\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.125875 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-config\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.132148 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-users\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.132507 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.132609 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.133003 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.133494 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.154407 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jghc6\" (UniqueName: \"kubernetes.io/projected/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-kube-api-access-jghc6\") pod \"default-interconnect-55bf8d5cb-rlct4\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.223404 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:26:44 crc kubenswrapper[5117]: I0321 00:26:44.493181 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-rlct4"] Mar 21 00:26:45 crc kubenswrapper[5117]: I0321 00:26:45.386813 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" event={"ID":"8034ad70-a7b5-4e15-91b6-a53fcbc9839a","Type":"ContainerStarted","Data":"a1a1ca6bb459a615017cccc5d48feedc976fe726db3aecbf3f103ba8cc5b5225"} Mar 21 00:26:50 crc kubenswrapper[5117]: I0321 00:26:50.430917 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" event={"ID":"8034ad70-a7b5-4e15-91b6-a53fcbc9839a","Type":"ContainerStarted","Data":"91afebddf6fa2d5a89cb3d8c574063185d82c6232fe90110641bc1a6d1c6c60c"} Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.400511 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" podStartSLOduration=6.388404586 podStartE2EDuration="11.400487765s" podCreationTimestamp="2026-03-21 00:26:43 +0000 UTC" firstStartedPulling="2026-03-21 00:26:44.517591193 +0000 UTC m=+917.810878395" lastFinishedPulling="2026-03-21 00:26:49.529674402 +0000 UTC m=+922.822961574" observedRunningTime="2026-03-21 00:26:50.454696867 +0000 UTC m=+923.747984079" watchObservedRunningTime="2026-03-21 00:26:54.400487765 +0000 UTC m=+927.693774947" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.408666 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.419430 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.419786 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.422586 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-default\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.422775 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-stf-dockercfg-2t4ql\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.422815 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-default-rulefiles-2\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.422832 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-default-rulefiles-1\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.422909 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-default-rulefiles-0\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.423049 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-default-web-config\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.423123 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-session-secret\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.425433 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-default-tls-assets-0\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.425596 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"serving-certs-ca-bundle\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.426269 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-prometheus-proxy-tls\"" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495019 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-config\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495068 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495102 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-12c615fb-a1df-43f4-9472-c4446e6c1f24\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12c615fb-a1df-43f4-9472-c4446e6c1f24\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495171 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-web-config\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495393 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b1f577e0-5a48-4c25-bbe2-7d20de72553f-tls-assets\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495445 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495608 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b1f577e0-5a48-4c25-bbe2-7d20de72553f-config-out\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495706 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495782 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpq69\" (UniqueName: \"kubernetes.io/projected/b1f577e0-5a48-4c25-bbe2-7d20de72553f-kube-api-access-vpq69\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495901 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.495982 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.496011 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597055 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b1f577e0-5a48-4c25-bbe2-7d20de72553f-tls-assets\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597142 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597191 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b1f577e0-5a48-4c25-bbe2-7d20de72553f-config-out\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597222 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597278 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vpq69\" (UniqueName: \"kubernetes.io/projected/b1f577e0-5a48-4c25-bbe2-7d20de72553f-kube-api-access-vpq69\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597330 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597412 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597451 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597498 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-config\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597529 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597596 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-12c615fb-a1df-43f4-9472-c4446e6c1f24\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12c615fb-a1df-43f4-9472-c4446e6c1f24\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.597624 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-web-config\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.598435 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: E0321 00:26:54.598782 5117 secret.go:189] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 21 00:26:54 crc kubenswrapper[5117]: E0321 00:26:54.598894 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-prometheus-proxy-tls podName:b1f577e0-5a48-4c25-bbe2-7d20de72553f nodeName:}" failed. No retries permitted until 2026-03-21 00:26:55.098862869 +0000 UTC m=+928.392150071 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "b1f577e0-5a48-4c25-bbe2-7d20de72553f") : secret "default-prometheus-proxy-tls" not found Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.599084 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.599629 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.602172 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1f577e0-5a48-4c25-bbe2-7d20de72553f-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.606463 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-config\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.606471 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b1f577e0-5a48-4c25-bbe2-7d20de72553f-tls-assets\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.608669 5117 csi_attacher.go:373] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.608709 5117 operation_generator.go:557] "MountVolume.MountDevice succeeded for volume \"pvc-12c615fb-a1df-43f4-9472-c4446e6c1f24\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12c615fb-a1df-43f4-9472-c4446e6c1f24\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a984a6e22c02253778b087538c6f08d0352a8ee8b865e1a8847fdb7a3785c1b6/globalmount\"" pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.611048 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b1f577e0-5a48-4c25-bbe2-7d20de72553f-config-out\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.611150 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-web-config\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.621770 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.627934 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpq69\" (UniqueName: \"kubernetes.io/projected/b1f577e0-5a48-4c25-bbe2-7d20de72553f-kube-api-access-vpq69\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.649201 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-12c615fb-a1df-43f4-9472-c4446e6c1f24\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12c615fb-a1df-43f4-9472-c4446e6c1f24\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.718694 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.718819 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.718891 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.719806 5117 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"324331d7c0a640c4e69a7d08e2bb3a2ebe5299eeda584b902a33658b2293b76e"} pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 00:26:54 crc kubenswrapper[5117]: I0321 00:26:54.719919 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" containerID="cri-o://324331d7c0a640c4e69a7d08e2bb3a2ebe5299eeda584b902a33658b2293b76e" gracePeriod=600 Mar 21 00:26:55 crc kubenswrapper[5117]: I0321 00:26:55.108973 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:55 crc kubenswrapper[5117]: E0321 00:26:55.109230 5117 secret.go:189] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 21 00:26:55 crc kubenswrapper[5117]: E0321 00:26:55.109374 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-prometheus-proxy-tls podName:b1f577e0-5a48-4c25-bbe2-7d20de72553f nodeName:}" failed. No retries permitted until 2026-03-21 00:26:56.109344604 +0000 UTC m=+929.402631816 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "b1f577e0-5a48-4c25-bbe2-7d20de72553f") : secret "default-prometheus-proxy-tls" not found Mar 21 00:26:55 crc kubenswrapper[5117]: I0321 00:26:55.470553 5117 generic.go:358] "Generic (PLEG): container finished" podID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerID="324331d7c0a640c4e69a7d08e2bb3a2ebe5299eeda584b902a33658b2293b76e" exitCode=0 Mar 21 00:26:55 crc kubenswrapper[5117]: I0321 00:26:55.470933 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerDied","Data":"324331d7c0a640c4e69a7d08e2bb3a2ebe5299eeda584b902a33658b2293b76e"} Mar 21 00:26:55 crc kubenswrapper[5117]: I0321 00:26:55.470958 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerStarted","Data":"eb586c1751d61bfe9aa828e1aeabf26cbd66967a6d89d22ac2c57498df26a314"} Mar 21 00:26:55 crc kubenswrapper[5117]: I0321 00:26:55.470975 5117 scope.go:117] "RemoveContainer" containerID="f2fd8729d8670a753e93005428f17279031e73d4ffabcffd4b6e8f29d6bcce8b" Mar 21 00:26:56 crc kubenswrapper[5117]: I0321 00:26:56.123738 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:56 crc kubenswrapper[5117]: I0321 00:26:56.133732 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1f577e0-5a48-4c25-bbe2-7d20de72553f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"b1f577e0-5a48-4c25-bbe2-7d20de72553f\") " pod="service-telemetry/prometheus-default-0" Mar 21 00:26:56 crc kubenswrapper[5117]: I0321 00:26:56.253012 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 21 00:26:56 crc kubenswrapper[5117]: I0321 00:26:56.539103 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 21 00:26:56 crc kubenswrapper[5117]: W0321 00:26:56.556224 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1f577e0_5a48_4c25_bbe2_7d20de72553f.slice/crio-74c417908efd6b77685e1d8d717b03b39f4c51153a97875ba461cc3de3f51240 WatchSource:0}: Error finding container 74c417908efd6b77685e1d8d717b03b39f4c51153a97875ba461cc3de3f51240: Status 404 returned error can't find the container with id 74c417908efd6b77685e1d8d717b03b39f4c51153a97875ba461cc3de3f51240 Mar 21 00:26:57 crc kubenswrapper[5117]: I0321 00:26:57.491699 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"b1f577e0-5a48-4c25-bbe2-7d20de72553f","Type":"ContainerStarted","Data":"74c417908efd6b77685e1d8d717b03b39f4c51153a97875ba461cc3de3f51240"} Mar 21 00:27:02 crc kubenswrapper[5117]: I0321 00:27:02.540432 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"b1f577e0-5a48-4c25-bbe2-7d20de72553f","Type":"ContainerStarted","Data":"2a6cbd51cf9c56cf19368fe059f92a64648a59d9fc4f0a788d151bde73360dff"} Mar 21 00:27:05 crc kubenswrapper[5117]: I0321 00:27:05.330153 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs"] Mar 21 00:27:05 crc kubenswrapper[5117]: I0321 00:27:05.366693 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs"] Mar 21 00:27:05 crc kubenswrapper[5117]: I0321 00:27:05.366864 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs" Mar 21 00:27:05 crc kubenswrapper[5117]: I0321 00:27:05.558490 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-727cf\" (UniqueName: \"kubernetes.io/projected/b4dbbbf2-a573-41be-b77c-4beba9161a98-kube-api-access-727cf\") pod \"default-snmp-webhook-6774d8dfbc-9mdbs\" (UID: \"b4dbbbf2-a573-41be-b77c-4beba9161a98\") " pod="service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs" Mar 21 00:27:05 crc kubenswrapper[5117]: I0321 00:27:05.659792 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-727cf\" (UniqueName: \"kubernetes.io/projected/b4dbbbf2-a573-41be-b77c-4beba9161a98-kube-api-access-727cf\") pod \"default-snmp-webhook-6774d8dfbc-9mdbs\" (UID: \"b4dbbbf2-a573-41be-b77c-4beba9161a98\") " pod="service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs" Mar 21 00:27:05 crc kubenswrapper[5117]: I0321 00:27:05.686412 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-727cf\" (UniqueName: \"kubernetes.io/projected/b4dbbbf2-a573-41be-b77c-4beba9161a98-kube-api-access-727cf\") pod \"default-snmp-webhook-6774d8dfbc-9mdbs\" (UID: \"b4dbbbf2-a573-41be-b77c-4beba9161a98\") " pod="service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs" Mar 21 00:27:05 crc kubenswrapper[5117]: I0321 00:27:05.986309 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs" Mar 21 00:27:06 crc kubenswrapper[5117]: I0321 00:27:06.249764 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs"] Mar 21 00:27:06 crc kubenswrapper[5117]: I0321 00:27:06.584900 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs" event={"ID":"b4dbbbf2-a573-41be-b77c-4beba9161a98","Type":"ContainerStarted","Data":"f380700857da5f879ecce0beda148dd6ed86f4e9afd7de3bf65a32a4add05030"} Mar 21 00:27:08 crc kubenswrapper[5117]: I0321 00:27:08.613081 5117 generic.go:358] "Generic (PLEG): container finished" podID="b1f577e0-5a48-4c25-bbe2-7d20de72553f" containerID="2a6cbd51cf9c56cf19368fe059f92a64648a59d9fc4f0a788d151bde73360dff" exitCode=0 Mar 21 00:27:08 crc kubenswrapper[5117]: I0321 00:27:08.613190 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"b1f577e0-5a48-4c25-bbe2-7d20de72553f","Type":"ContainerDied","Data":"2a6cbd51cf9c56cf19368fe059f92a64648a59d9fc4f0a788d151bde73360dff"} Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.251081 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.265651 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.272208 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-cluster-tls-config\"" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.272453 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-alertmanager-proxy-tls\"" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.272514 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-generated\"" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.272598 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-tls-assets-0\"" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.273636 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-stf-dockercfg-59nk7\"" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.285055 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-web-config\"" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.300292 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.365187 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a4667f7e-eb34-4875-9bf2-869283b1c7c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a4667f7e-eb34-4875-9bf2-869283b1c7c8\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.365530 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-config-volume\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.365559 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2tzx\" (UniqueName: \"kubernetes.io/projected/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-kube-api-access-l2tzx\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.365584 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.365607 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-tls-assets\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.365652 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-web-config\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.365681 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-config-out\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.365704 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.365720 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.467031 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-config-out\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.467084 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.467106 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.467149 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-a4667f7e-eb34-4875-9bf2-869283b1c7c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a4667f7e-eb34-4875-9bf2-869283b1c7c8\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.467171 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-config-volume\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.467194 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l2tzx\" (UniqueName: \"kubernetes.io/projected/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-kube-api-access-l2tzx\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.467215 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.467238 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-tls-assets\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.467297 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-web-config\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.480684 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-config-out\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.480985 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-web-config\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.480985 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.480988 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.480990 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-config-volume\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.481215 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.482466 5117 csi_attacher.go:373] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.482499 5117 operation_generator.go:557] "MountVolume.MountDevice succeeded for volume \"pvc-a4667f7e-eb34-4875-9bf2-869283b1c7c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a4667f7e-eb34-4875-9bf2-869283b1c7c8\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/904f65ab7a0b684cbe350b86c384628384e8e5a2763e72c6234639e7bf9e0d3b/globalmount\"" pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.495953 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2tzx\" (UniqueName: \"kubernetes.io/projected/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-kube-api-access-l2tzx\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.496857 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/cd169b92-c7d5-4c3a-a6ce-db74aa6422ed-tls-assets\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.508576 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-a4667f7e-eb34-4875-9bf2-869283b1c7c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a4667f7e-eb34-4875-9bf2-869283b1c7c8\") pod \"alertmanager-default-0\" (UID: \"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed\") " pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:13 crc kubenswrapper[5117]: I0321 00:27:13.593819 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 21 00:27:15 crc kubenswrapper[5117]: I0321 00:27:15.016041 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 21 00:27:15 crc kubenswrapper[5117]: I0321 00:27:15.684743 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs" event={"ID":"b4dbbbf2-a573-41be-b77c-4beba9161a98","Type":"ContainerStarted","Data":"2eafc6a66e101888fcdcb62096a53942bf3bdbf7673f306d799d058ae11776e5"} Mar 21 00:27:15 crc kubenswrapper[5117]: I0321 00:27:15.693722 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed","Type":"ContainerStarted","Data":"7498d19dca16eec937dcf88b857b11b15eb66e9449359661b8f493e012b8141d"} Mar 21 00:27:15 crc kubenswrapper[5117]: I0321 00:27:15.704663 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-9mdbs" podStartSLOduration=2.112839733 podStartE2EDuration="10.704649061s" podCreationTimestamp="2026-03-21 00:27:05 +0000 UTC" firstStartedPulling="2026-03-21 00:27:06.269648129 +0000 UTC m=+939.562935301" lastFinishedPulling="2026-03-21 00:27:14.861457467 +0000 UTC m=+948.154744629" observedRunningTime="2026-03-21 00:27:15.702133193 +0000 UTC m=+948.995420375" watchObservedRunningTime="2026-03-21 00:27:15.704649061 +0000 UTC m=+948.997936233" Mar 21 00:27:17 crc kubenswrapper[5117]: I0321 00:27:17.711899 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed","Type":"ContainerStarted","Data":"0bfc72928f424e4e809ca3d8f58b6b04364e43bf7ce637c82e6a4c1f0723a467"} Mar 21 00:27:20 crc kubenswrapper[5117]: I0321 00:27:20.768352 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"b1f577e0-5a48-4c25-bbe2-7d20de72553f","Type":"ContainerStarted","Data":"b2256acddd3f3b322784ce477ec10556ac2aa09be1101254b98fccf15556c544"} Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.227831 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj"] Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.239561 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.243702 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"smart-gateway-session-secret\"" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.243944 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-cloud1-coll-meter-proxy-tls\"" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.248772 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-coll-meter-sg-core-configmap\"" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.248849 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"smart-gateway-dockercfg-w298h\"" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.250729 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj"] Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.326611 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.326693 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/cd862294-3d4d-42a8-ba79-fc7607eac35e-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.326734 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5cnc\" (UniqueName: \"kubernetes.io/projected/cd862294-3d4d-42a8-ba79-fc7607eac35e-kube-api-access-p5cnc\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.326925 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.327017 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/cd862294-3d4d-42a8-ba79-fc7607eac35e-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.428208 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.428324 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/cd862294-3d4d-42a8-ba79-fc7607eac35e-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.428359 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p5cnc\" (UniqueName: \"kubernetes.io/projected/cd862294-3d4d-42a8-ba79-fc7607eac35e-kube-api-access-p5cnc\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: E0321 00:27:22.428375 5117 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.428407 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: E0321 00:27:22.428461 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-default-cloud1-coll-meter-proxy-tls podName:cd862294-3d4d-42a8-ba79-fc7607eac35e nodeName:}" failed. No retries permitted until 2026-03-21 00:27:22.928438153 +0000 UTC m=+956.221725325 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" (UID: "cd862294-3d4d-42a8-ba79-fc7607eac35e") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.428496 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/cd862294-3d4d-42a8-ba79-fc7607eac35e-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.429013 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/cd862294-3d4d-42a8-ba79-fc7607eac35e-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.429333 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/cd862294-3d4d-42a8-ba79-fc7607eac35e-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.444511 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.454383 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5cnc\" (UniqueName: \"kubernetes.io/projected/cd862294-3d4d-42a8-ba79-fc7607eac35e-kube-api-access-p5cnc\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: I0321 00:27:22.935938 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:22 crc kubenswrapper[5117]: E0321 00:27:22.936820 5117 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 21 00:27:22 crc kubenswrapper[5117]: E0321 00:27:22.937053 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-default-cloud1-coll-meter-proxy-tls podName:cd862294-3d4d-42a8-ba79-fc7607eac35e nodeName:}" failed. No retries permitted until 2026-03-21 00:27:23.93702595 +0000 UTC m=+957.230313152 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" (UID: "cd862294-3d4d-42a8-ba79-fc7607eac35e") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 21 00:27:23 crc kubenswrapper[5117]: I0321 00:27:23.790920 5117 generic.go:358] "Generic (PLEG): container finished" podID="cd169b92-c7d5-4c3a-a6ce-db74aa6422ed" containerID="0bfc72928f424e4e809ca3d8f58b6b04364e43bf7ce637c82e6a4c1f0723a467" exitCode=0 Mar 21 00:27:23 crc kubenswrapper[5117]: I0321 00:27:23.791076 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed","Type":"ContainerDied","Data":"0bfc72928f424e4e809ca3d8f58b6b04364e43bf7ce637c82e6a4c1f0723a467"} Mar 21 00:27:23 crc kubenswrapper[5117]: I0321 00:27:23.793163 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"b1f577e0-5a48-4c25-bbe2-7d20de72553f","Type":"ContainerStarted","Data":"ae8855f1da10fbebaa9ae5b7da0d37ae03dde61245d02c3d484073c49a096f18"} Mar 21 00:27:23 crc kubenswrapper[5117]: I0321 00:27:23.953207 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:23 crc kubenswrapper[5117]: I0321 00:27:23.962380 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd862294-3d4d-42a8-ba79-fc7607eac35e-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-qvvvj\" (UID: \"cd862294-3d4d-42a8-ba79-fc7607eac35e\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.058663 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.564151 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj"] Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.888322 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289"] Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.896312 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.899958 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289"] Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.901748 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-ceil-meter-sg-core-configmap\"" Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.901808 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-cloud1-ceil-meter-proxy-tls\"" Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.970604 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.970673 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.970734 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.970981 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:24 crc kubenswrapper[5117]: I0321 00:27:24.971018 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz8mg\" (UniqueName: \"kubernetes.io/projected/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-kube-api-access-mz8mg\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.072046 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.072096 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.072134 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.072154 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.072189 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mz8mg\" (UniqueName: \"kubernetes.io/projected/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-kube-api-access-mz8mg\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: E0321 00:27:25.073610 5117 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 21 00:27:25 crc kubenswrapper[5117]: E0321 00:27:25.073665 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-default-cloud1-ceil-meter-proxy-tls podName:e6c95eaa-fc0b-48f0-93bf-1d2deefc2362 nodeName:}" failed. No retries permitted until 2026-03-21 00:27:25.573649162 +0000 UTC m=+958.866936334 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" (UID: "e6c95eaa-fc0b-48f0-93bf-1d2deefc2362") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.074126 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.074576 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.094933 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz8mg\" (UniqueName: \"kubernetes.io/projected/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-kube-api-access-mz8mg\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.095756 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.579209 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:25 crc kubenswrapper[5117]: E0321 00:27:25.579340 5117 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 21 00:27:25 crc kubenswrapper[5117]: E0321 00:27:25.579809 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-default-cloud1-ceil-meter-proxy-tls podName:e6c95eaa-fc0b-48f0-93bf-1d2deefc2362 nodeName:}" failed. No retries permitted until 2026-03-21 00:27:26.579790563 +0000 UTC m=+959.873077735 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" (UID: "e6c95eaa-fc0b-48f0-93bf-1d2deefc2362") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 21 00:27:25 crc kubenswrapper[5117]: I0321 00:27:25.840865 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" event={"ID":"cd862294-3d4d-42a8-ba79-fc7607eac35e","Type":"ContainerStarted","Data":"aed24723cfc73151b0df96707b32c19eaef49f470f5839ede8f0b88066af75d0"} Mar 21 00:27:26 crc kubenswrapper[5117]: I0321 00:27:26.598912 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:26 crc kubenswrapper[5117]: I0321 00:27:26.605674 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e6c95eaa-fc0b-48f0-93bf-1d2deefc2362-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289\" (UID: \"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:26 crc kubenswrapper[5117]: I0321 00:27:26.720539 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.707955 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth"] Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.727667 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth"] Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.727807 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.732102 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-cloud1-sens-meter-proxy-tls\"" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.732324 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-sens-meter-sg-core-configmap\"" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.835912 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwk5r\" (UniqueName: \"kubernetes.io/projected/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-kube-api-access-pwk5r\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.836201 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.836267 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.836289 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.836311 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.937778 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pwk5r\" (UniqueName: \"kubernetes.io/projected/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-kube-api-access-pwk5r\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.937860 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.937922 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.937939 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.937973 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: E0321 00:27:28.938142 5117 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 21 00:27:28 crc kubenswrapper[5117]: E0321 00:27:28.938262 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-default-cloud1-sens-meter-proxy-tls podName:cafef1e9-9a8d-4e86-b8ac-8480b2f141b4 nodeName:}" failed. No retries permitted until 2026-03-21 00:27:29.438222108 +0000 UTC m=+962.731509280 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" (UID: "cafef1e9-9a8d-4e86-b8ac-8480b2f141b4") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.938635 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.938895 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.955292 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:28 crc kubenswrapper[5117]: I0321 00:27:28.957787 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwk5r\" (UniqueName: \"kubernetes.io/projected/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-kube-api-access-pwk5r\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:29 crc kubenswrapper[5117]: I0321 00:27:29.447147 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:29 crc kubenswrapper[5117]: E0321 00:27:29.447368 5117 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 21 00:27:29 crc kubenswrapper[5117]: E0321 00:27:29.447420 5117 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-default-cloud1-sens-meter-proxy-tls podName:cafef1e9-9a8d-4e86-b8ac-8480b2f141b4 nodeName:}" failed. No retries permitted until 2026-03-21 00:27:30.447404653 +0000 UTC m=+963.740691825 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" (UID: "cafef1e9-9a8d-4e86-b8ac-8480b2f141b4") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 21 00:27:30 crc kubenswrapper[5117]: I0321 00:27:30.464711 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:30 crc kubenswrapper[5117]: I0321 00:27:30.491719 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/cafef1e9-9a8d-4e86-b8ac-8480b2f141b4-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth\" (UID: \"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:30 crc kubenswrapper[5117]: I0321 00:27:30.553440 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" Mar 21 00:27:30 crc kubenswrapper[5117]: W0321 00:27:30.958403 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6c95eaa_fc0b_48f0_93bf_1d2deefc2362.slice/crio-a63ce13efa137830b2d6c66a7ff152ae18249537ef7248694cc89ec60273e8b7 WatchSource:0}: Error finding container a63ce13efa137830b2d6c66a7ff152ae18249537ef7248694cc89ec60273e8b7: Status 404 returned error can't find the container with id a63ce13efa137830b2d6c66a7ff152ae18249537ef7248694cc89ec60273e8b7 Mar 21 00:27:30 crc kubenswrapper[5117]: I0321 00:27:30.967869 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289"] Mar 21 00:27:31 crc kubenswrapper[5117]: I0321 00:27:31.227016 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth"] Mar 21 00:27:31 crc kubenswrapper[5117]: I0321 00:27:31.898488 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed","Type":"ContainerStarted","Data":"67cb4773cc96f72827e0a98109b31020b42f141ff8599fd7c354f289bdc96e1d"} Mar 21 00:27:31 crc kubenswrapper[5117]: I0321 00:27:31.899814 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" event={"ID":"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4","Type":"ContainerStarted","Data":"cbfeb16647f1cfcb640cfe8a5ddb646c281eeabeb9e9c9486a7ddbb69e60e808"} Mar 21 00:27:31 crc kubenswrapper[5117]: I0321 00:27:31.903047 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"b1f577e0-5a48-4c25-bbe2-7d20de72553f","Type":"ContainerStarted","Data":"e0c74bfc388f0bdaaae3ad0ff10199dea70aea4a234ef9b45f674fa225480549"} Mar 21 00:27:31 crc kubenswrapper[5117]: I0321 00:27:31.904955 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" event={"ID":"cd862294-3d4d-42a8-ba79-fc7607eac35e","Type":"ContainerStarted","Data":"8a64ad5095d5504bcde568b69db9cc2c9aee16d9f34a6e12fc8b3385e3ba31be"} Mar 21 00:27:31 crc kubenswrapper[5117]: I0321 00:27:31.912436 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" event={"ID":"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362","Type":"ContainerStarted","Data":"a63ce13efa137830b2d6c66a7ff152ae18249537ef7248694cc89ec60273e8b7"} Mar 21 00:27:31 crc kubenswrapper[5117]: I0321 00:27:31.914722 5117 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 00:27:31 crc kubenswrapper[5117]: I0321 00:27:31.932842 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=4.724200855 podStartE2EDuration="38.932826311s" podCreationTimestamp="2026-03-21 00:26:53 +0000 UTC" firstStartedPulling="2026-03-21 00:26:56.562317615 +0000 UTC m=+929.855604787" lastFinishedPulling="2026-03-21 00:27:30.770943071 +0000 UTC m=+964.064230243" observedRunningTime="2026-03-21 00:27:31.931520021 +0000 UTC m=+965.224807193" watchObservedRunningTime="2026-03-21 00:27:31.932826311 +0000 UTC m=+965.226113483" Mar 21 00:27:32 crc kubenswrapper[5117]: I0321 00:27:32.921547 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed","Type":"ContainerStarted","Data":"3780a5cc99ea0748bd07fbef618ce0ceeb2f54dc900b6428a8e3fb32ab69b5f4"} Mar 21 00:27:32 crc kubenswrapper[5117]: I0321 00:27:32.923483 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" event={"ID":"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4","Type":"ContainerStarted","Data":"950049a2be4fd444bbb7cc5baf154b3f482bb193fd507609aed69390b63da199"} Mar 21 00:27:32 crc kubenswrapper[5117]: I0321 00:27:32.925150 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" event={"ID":"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362","Type":"ContainerStarted","Data":"6421d1518459f01a399ebc39daa2a1ab98b1d1b49ca997b6ffd6515817ff29ca"} Mar 21 00:27:33 crc kubenswrapper[5117]: I0321 00:27:33.956148 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"cd169b92-c7d5-4c3a-a6ce-db74aa6422ed","Type":"ContainerStarted","Data":"2f5377d354e8fc0c9ff337e253674e923bb51c870460df5d01a24a9cccdaa300"} Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.253908 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="service-telemetry/prometheus-default-0" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.410334 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=14.871763996 podStartE2EDuration="24.41031399s" podCreationTimestamp="2026-03-21 00:27:12 +0000 UTC" firstStartedPulling="2026-03-21 00:27:23.792793714 +0000 UTC m=+957.086080896" lastFinishedPulling="2026-03-21 00:27:33.331343718 +0000 UTC m=+966.624630890" observedRunningTime="2026-03-21 00:27:33.97883508 +0000 UTC m=+967.272122252" watchObservedRunningTime="2026-03-21 00:27:36.41031399 +0000 UTC m=+969.703601162" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.412834 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c"] Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.422576 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c"] Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.422705 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.424998 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-coll-event-sg-core-configmap\"" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.431614 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-cert\"" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.562497 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.562579 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.562604 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n4lm\" (UniqueName: \"kubernetes.io/projected/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-kube-api-access-9n4lm\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.562637 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.664318 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.664383 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.664410 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9n4lm\" (UniqueName: \"kubernetes.io/projected/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-kube-api-access-9n4lm\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.664452 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.665437 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.665478 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.670469 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.683922 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n4lm\" (UniqueName: \"kubernetes.io/projected/b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb-kube-api-access-9n4lm\") pod \"default-cloud1-coll-event-smartgateway-6f48c45984-chc9c\" (UID: \"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:36 crc kubenswrapper[5117]: I0321 00:27:36.754947 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.122630 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6"] Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.145626 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6"] Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.145805 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.148656 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-ceil-event-sg-core-configmap\"" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.273917 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e1e39c35-dd28-4061-b744-debef1ae6e42-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.273976 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75b79\" (UniqueName: \"kubernetes.io/projected/e1e39c35-dd28-4061-b744-debef1ae6e42-kube-api-access-75b79\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.274221 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/e1e39c35-dd28-4061-b744-debef1ae6e42-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.274381 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e1e39c35-dd28-4061-b744-debef1ae6e42-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.375768 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e1e39c35-dd28-4061-b744-debef1ae6e42-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.375868 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-75b79\" (UniqueName: \"kubernetes.io/projected/e1e39c35-dd28-4061-b744-debef1ae6e42-kube-api-access-75b79\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.376327 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/e1e39c35-dd28-4061-b744-debef1ae6e42-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.376405 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e1e39c35-dd28-4061-b744-debef1ae6e42-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.376714 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e1e39c35-dd28-4061-b744-debef1ae6e42-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.376833 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e1e39c35-dd28-4061-b744-debef1ae6e42-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.384831 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/e1e39c35-dd28-4061-b744-debef1ae6e42-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.402011 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-75b79\" (UniqueName: \"kubernetes.io/projected/e1e39c35-dd28-4061-b744-debef1ae6e42-kube-api-access-75b79\") pod \"default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6\" (UID: \"e1e39c35-dd28-4061-b744-debef1ae6e42\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:37 crc kubenswrapper[5117]: I0321 00:27:37.461510 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" Mar 21 00:27:39 crc kubenswrapper[5117]: I0321 00:27:39.038506 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6"] Mar 21 00:27:39 crc kubenswrapper[5117]: I0321 00:27:39.173334 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c"] Mar 21 00:27:39 crc kubenswrapper[5117]: W0321 00:27:39.175204 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1c9e2b8_36c4_495c_95d5_a04cbdd00bdb.slice/crio-304f1ca8579d1b9fb572be55ef9ba28ecbbc6431b48e0717b15694f245400d02 WatchSource:0}: Error finding container 304f1ca8579d1b9fb572be55ef9ba28ecbbc6431b48e0717b15694f245400d02: Status 404 returned error can't find the container with id 304f1ca8579d1b9fb572be55ef9ba28ecbbc6431b48e0717b15694f245400d02 Mar 21 00:27:40 crc kubenswrapper[5117]: I0321 00:27:40.007615 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" event={"ID":"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb","Type":"ContainerStarted","Data":"304f1ca8579d1b9fb572be55ef9ba28ecbbc6431b48e0717b15694f245400d02"} Mar 21 00:27:40 crc kubenswrapper[5117]: I0321 00:27:40.016898 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" event={"ID":"cd862294-3d4d-42a8-ba79-fc7607eac35e","Type":"ContainerStarted","Data":"d9853fe504cf254d5533cdd3bbf732817f175fe6ee0783c4e75808f93676231b"} Mar 21 00:27:40 crc kubenswrapper[5117]: I0321 00:27:40.019098 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" event={"ID":"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362","Type":"ContainerStarted","Data":"f13f8c346488a4dc3ee3d945c33a582b1891e2d56b654c0e3d56f98f2eae6770"} Mar 21 00:27:40 crc kubenswrapper[5117]: I0321 00:27:40.021282 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" event={"ID":"e1e39c35-dd28-4061-b744-debef1ae6e42","Type":"ContainerStarted","Data":"e03f1f1ef3145c8c131d7c8206dbdd1984572262516743ed38b5f8fb6a584e0d"} Mar 21 00:27:40 crc kubenswrapper[5117]: I0321 00:27:40.021326 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" event={"ID":"e1e39c35-dd28-4061-b744-debef1ae6e42","Type":"ContainerStarted","Data":"f384058a6002cf6e8aa95d14e7e842bcf2403ad88884a0775654fcb5fa89dee7"} Mar 21 00:27:40 crc kubenswrapper[5117]: I0321 00:27:40.023223 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" event={"ID":"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4","Type":"ContainerStarted","Data":"404cb8b77731126a7b75429f875177e1803b8d6d318b220aed5b23959261ef59"} Mar 21 00:27:41 crc kubenswrapper[5117]: I0321 00:27:41.045769 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" event={"ID":"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb","Type":"ContainerStarted","Data":"5ef900960997b2d513ec5718872126bd8e3f6b7f92518666bd8f01756f52cd50"} Mar 21 00:27:41 crc kubenswrapper[5117]: I0321 00:27:41.253278 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Mar 21 00:27:41 crc kubenswrapper[5117]: I0321 00:27:41.301715 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Mar 21 00:27:42 crc kubenswrapper[5117]: I0321 00:27:42.091399 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Mar 21 00:27:49 crc kubenswrapper[5117]: I0321 00:27:49.594315 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-rlct4"] Mar 21 00:27:49 crc kubenswrapper[5117]: I0321 00:27:49.594902 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" podUID="8034ad70-a7b5-4e15-91b6-a53fcbc9839a" containerName="default-interconnect" containerID="cri-o://91afebddf6fa2d5a89cb3d8c574063185d82c6232fe90110641bc1a6d1c6c60c" gracePeriod=30 Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.123643 5117 generic.go:358] "Generic (PLEG): container finished" podID="8034ad70-a7b5-4e15-91b6-a53fcbc9839a" containerID="91afebddf6fa2d5a89cb3d8c574063185d82c6232fe90110641bc1a6d1c6c60c" exitCode=0 Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.123773 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" event={"ID":"8034ad70-a7b5-4e15-91b6-a53fcbc9839a","Type":"ContainerDied","Data":"91afebddf6fa2d5a89cb3d8c574063185d82c6232fe90110641bc1a6d1c6c60c"} Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.127936 5117 generic.go:358] "Generic (PLEG): container finished" podID="e1e39c35-dd28-4061-b744-debef1ae6e42" containerID="e03f1f1ef3145c8c131d7c8206dbdd1984572262516743ed38b5f8fb6a584e0d" exitCode=0 Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.127950 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" event={"ID":"e1e39c35-dd28-4061-b744-debef1ae6e42","Type":"ContainerDied","Data":"e03f1f1ef3145c8c131d7c8206dbdd1984572262516743ed38b5f8fb6a584e0d"} Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.404285 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.449017 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-x4qfv"] Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.449805 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8034ad70-a7b5-4e15-91b6-a53fcbc9839a" containerName="default-interconnect" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.449826 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="8034ad70-a7b5-4e15-91b6-a53fcbc9839a" containerName="default-interconnect" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.449951 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="8034ad70-a7b5-4e15-91b6-a53fcbc9839a" containerName="default-interconnect" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.455745 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-x4qfv"] Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.455891 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.490351 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jghc6\" (UniqueName: \"kubernetes.io/projected/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-kube-api-access-jghc6\") pod \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.490420 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-ca\") pod \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.490525 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-users\") pod \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.490567 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-credentials\") pod \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.490609 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-config\") pod \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.490896 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-ca\") pod \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.490943 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-credentials\") pod \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\" (UID: \"8034ad70-a7b5-4e15-91b6-a53fcbc9839a\") " Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.496201 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-kube-api-access-jghc6" (OuterVolumeSpecName: "kube-api-access-jghc6") pod "8034ad70-a7b5-4e15-91b6-a53fcbc9839a" (UID: "8034ad70-a7b5-4e15-91b6-a53fcbc9839a"). InnerVolumeSpecName "kube-api-access-jghc6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.502545 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "8034ad70-a7b5-4e15-91b6-a53fcbc9839a" (UID: "8034ad70-a7b5-4e15-91b6-a53fcbc9839a"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.504598 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "8034ad70-a7b5-4e15-91b6-a53fcbc9839a" (UID: "8034ad70-a7b5-4e15-91b6-a53fcbc9839a"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.507242 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "8034ad70-a7b5-4e15-91b6-a53fcbc9839a" (UID: "8034ad70-a7b5-4e15-91b6-a53fcbc9839a"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.507309 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "8034ad70-a7b5-4e15-91b6-a53fcbc9839a" (UID: "8034ad70-a7b5-4e15-91b6-a53fcbc9839a"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.509479 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "8034ad70-a7b5-4e15-91b6-a53fcbc9839a" (UID: "8034ad70-a7b5-4e15-91b6-a53fcbc9839a"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.510750 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "8034ad70-a7b5-4e15-91b6-a53fcbc9839a" (UID: "8034ad70-a7b5-4e15-91b6-a53fcbc9839a"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.592752 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593130 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b62f948f-1381-4e90-a11c-e5ec089639ef-sasl-config\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593151 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pzxr\" (UniqueName: \"kubernetes.io/projected/b62f948f-1381-4e90-a11c-e5ec089639ef-kube-api-access-7pzxr\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593176 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593200 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593236 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-sasl-users\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593492 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593637 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-jghc6\" (UniqueName: \"kubernetes.io/projected/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-kube-api-access-jghc6\") on node \"crc\" DevicePath \"\"" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593653 5117 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593666 5117 reconciler_common.go:299] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-users\") on node \"crc\" DevicePath \"\"" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593677 5117 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593689 5117 reconciler_common.go:299] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-sasl-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593721 5117 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.593731 5117 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/8034ad70-a7b5-4e15-91b6-a53fcbc9839a-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.695315 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b62f948f-1381-4e90-a11c-e5ec089639ef-sasl-config\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.695401 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7pzxr\" (UniqueName: \"kubernetes.io/projected/b62f948f-1381-4e90-a11c-e5ec089639ef-kube-api-access-7pzxr\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.695478 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.695526 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.695627 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-sasl-users\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.695724 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.695832 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.697237 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b62f948f-1381-4e90-a11c-e5ec089639ef-sasl-config\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.700142 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-sasl-users\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.700157 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.700523 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.701056 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.702326 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b62f948f-1381-4e90-a11c-e5ec089639ef-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.725321 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pzxr\" (UniqueName: \"kubernetes.io/projected/b62f948f-1381-4e90-a11c-e5ec089639ef-kube-api-access-7pzxr\") pod \"default-interconnect-55bf8d5cb-x4qfv\" (UID: \"b62f948f-1381-4e90-a11c-e5ec089639ef\") " pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:50 crc kubenswrapper[5117]: I0321 00:27:50.799193 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.062510 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-x4qfv"] Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.143345 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" event={"ID":"b62f948f-1381-4e90-a11c-e5ec089639ef","Type":"ContainerStarted","Data":"bf7340f5e69298c1bc73690eb11b4a0e4095a6c0b03299f72aad67bfabf4df64"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.147993 5117 generic.go:358] "Generic (PLEG): container finished" podID="cafef1e9-9a8d-4e86-b8ac-8480b2f141b4" containerID="404cb8b77731126a7b75429f875177e1803b8d6d318b220aed5b23959261ef59" exitCode=0 Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.148063 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" event={"ID":"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4","Type":"ContainerDied","Data":"404cb8b77731126a7b75429f875177e1803b8d6d318b220aed5b23959261ef59"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.148109 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" event={"ID":"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4","Type":"ContainerStarted","Data":"915efefc36d103cd62ee3913e2195a5107db4ef1d4d8d147415cdd227bd3f8b6"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.148750 5117 scope.go:117] "RemoveContainer" containerID="404cb8b77731126a7b75429f875177e1803b8d6d318b220aed5b23959261ef59" Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.158037 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" event={"ID":"8034ad70-a7b5-4e15-91b6-a53fcbc9839a","Type":"ContainerDied","Data":"a1a1ca6bb459a615017cccc5d48feedc976fe726db3aecbf3f103ba8cc5b5225"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.158071 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-rlct4" Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.158351 5117 scope.go:117] "RemoveContainer" containerID="91afebddf6fa2d5a89cb3d8c574063185d82c6232fe90110641bc1a6d1c6c60c" Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.160435 5117 generic.go:358] "Generic (PLEG): container finished" podID="b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb" containerID="5ef900960997b2d513ec5718872126bd8e3f6b7f92518666bd8f01756f52cd50" exitCode=0 Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.160511 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" event={"ID":"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb","Type":"ContainerDied","Data":"5ef900960997b2d513ec5718872126bd8e3f6b7f92518666bd8f01756f52cd50"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.160546 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" event={"ID":"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb","Type":"ContainerStarted","Data":"bfdd5e23775e2a6dd7aeda0fddef1aa15cd68c9fe6ff73af910b18dc561a3c51"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.160804 5117 scope.go:117] "RemoveContainer" containerID="5ef900960997b2d513ec5718872126bd8e3f6b7f92518666bd8f01756f52cd50" Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.165888 5117 generic.go:358] "Generic (PLEG): container finished" podID="cd862294-3d4d-42a8-ba79-fc7607eac35e" containerID="d9853fe504cf254d5533cdd3bbf732817f175fe6ee0783c4e75808f93676231b" exitCode=0 Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.166062 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" event={"ID":"cd862294-3d4d-42a8-ba79-fc7607eac35e","Type":"ContainerDied","Data":"d9853fe504cf254d5533cdd3bbf732817f175fe6ee0783c4e75808f93676231b"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.166181 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" event={"ID":"cd862294-3d4d-42a8-ba79-fc7607eac35e","Type":"ContainerStarted","Data":"0c38304e976dc5f9e47f9c7b5a46dc7d44c9f5ce5bd2573fd5f01d904f82c906"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.166589 5117 scope.go:117] "RemoveContainer" containerID="d9853fe504cf254d5533cdd3bbf732817f175fe6ee0783c4e75808f93676231b" Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.175305 5117 generic.go:358] "Generic (PLEG): container finished" podID="e6c95eaa-fc0b-48f0-93bf-1d2deefc2362" containerID="f13f8c346488a4dc3ee3d945c33a582b1891e2d56b654c0e3d56f98f2eae6770" exitCode=0 Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.175591 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" event={"ID":"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362","Type":"ContainerDied","Data":"f13f8c346488a4dc3ee3d945c33a582b1891e2d56b654c0e3d56f98f2eae6770"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.175649 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" event={"ID":"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362","Type":"ContainerStarted","Data":"2cc13f57b363e8b88db8194136a20ba0794c72101d81d2e3bad32fdea65ebb3c"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.176894 5117 scope.go:117] "RemoveContainer" containerID="f13f8c346488a4dc3ee3d945c33a582b1891e2d56b654c0e3d56f98f2eae6770" Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.181497 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" event={"ID":"e1e39c35-dd28-4061-b744-debef1ae6e42","Type":"ContainerStarted","Data":"549713035e98959fd3fac0b527f3b8e7acb478abb57c30a34d70af7416a2d00d"} Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.181963 5117 scope.go:117] "RemoveContainer" containerID="e03f1f1ef3145c8c131d7c8206dbdd1984572262516743ed38b5f8fb6a584e0d" Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.265179 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-rlct4"] Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.271400 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-rlct4"] Mar 21 00:27:51 crc kubenswrapper[5117]: E0321 00:27:51.311765 5117 cadvisor_stats_provider.go:525] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8034ad70_a7b5_4e15_91b6_a53fcbc9839a.slice/crio-a1a1ca6bb459a615017cccc5d48feedc976fe726db3aecbf3f103ba8cc5b5225\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8034ad70_a7b5_4e15_91b6_a53fcbc9839a.slice\": RecentStats: unable to find data in memory cache]" Mar 21 00:27:51 crc kubenswrapper[5117]: I0321 00:27:51.587904 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8034ad70-a7b5-4e15-91b6-a53fcbc9839a" path="/var/lib/kubelet/pods/8034ad70-a7b5-4e15-91b6-a53fcbc9839a/volumes" Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.195130 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" event={"ID":"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4","Type":"ContainerStarted","Data":"256932e7300d587c1aab74c746b27af4ae2d4e1ca07e8defbe576b4d3c988349"} Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.202483 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" event={"ID":"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb","Type":"ContainerStarted","Data":"30f23160495446af3551d7451c46a512eeb7b95c5513bab2c653bf8272da4bda"} Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.207125 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" event={"ID":"cd862294-3d4d-42a8-ba79-fc7607eac35e","Type":"ContainerStarted","Data":"5ea72378e01ebc98f30b831b529fa2af979904fca63c30b0286543a641d41733"} Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.213505 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" event={"ID":"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362","Type":"ContainerStarted","Data":"44723ea181acb256f9a1045f7c471a47bcb50f9a4e68425eb2986a794f99582e"} Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.218285 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" event={"ID":"e1e39c35-dd28-4061-b744-debef1ae6e42","Type":"ContainerStarted","Data":"51ab96e2a99bbc9d35a9307ae47ab458e88a0f34986362e8afbfc113a3889a8b"} Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.221012 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" event={"ID":"b62f948f-1381-4e90-a11c-e5ec089639ef","Type":"ContainerStarted","Data":"b12a2473eea22751cc19ef6866c4536d6a93876b2ed515de815dbd782c37d2ef"} Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.238887 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" podStartSLOduration=3.810677564 podStartE2EDuration="24.238864963s" podCreationTimestamp="2026-03-21 00:27:28 +0000 UTC" firstStartedPulling="2026-03-21 00:27:31.235769435 +0000 UTC m=+964.529056607" lastFinishedPulling="2026-03-21 00:27:51.663956834 +0000 UTC m=+984.957244006" observedRunningTime="2026-03-21 00:27:52.232744293 +0000 UTC m=+985.526031505" watchObservedRunningTime="2026-03-21 00:27:52.238864963 +0000 UTC m=+985.532152175" Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.274336 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" podStartSLOduration=2.528175194 podStartE2EDuration="15.274298688s" podCreationTimestamp="2026-03-21 00:27:37 +0000 UTC" firstStartedPulling="2026-03-21 00:27:39.054752744 +0000 UTC m=+972.348039916" lastFinishedPulling="2026-03-21 00:27:51.800876238 +0000 UTC m=+985.094163410" observedRunningTime="2026-03-21 00:27:52.264504826 +0000 UTC m=+985.557792038" watchObservedRunningTime="2026-03-21 00:27:52.274298688 +0000 UTC m=+985.567585900" Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.296875 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-55bf8d5cb-x4qfv" podStartSLOduration=3.296847845 podStartE2EDuration="3.296847845s" podCreationTimestamp="2026-03-21 00:27:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-21 00:27:52.295232296 +0000 UTC m=+985.588519478" watchObservedRunningTime="2026-03-21 00:27:52.296847845 +0000 UTC m=+985.590135037" Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.324218 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" podStartSLOduration=3.697985219 podStartE2EDuration="30.324194001s" podCreationTimestamp="2026-03-21 00:27:22 +0000 UTC" firstStartedPulling="2026-03-21 00:27:24.987496348 +0000 UTC m=+958.280783520" lastFinishedPulling="2026-03-21 00:27:51.61370513 +0000 UTC m=+984.906992302" observedRunningTime="2026-03-21 00:27:52.322568231 +0000 UTC m=+985.615855423" watchObservedRunningTime="2026-03-21 00:27:52.324194001 +0000 UTC m=+985.617481173" Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.366696 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" podStartSLOduration=7.737422929 podStartE2EDuration="28.366678705s" podCreationTimestamp="2026-03-21 00:27:24 +0000 UTC" firstStartedPulling="2026-03-21 00:27:30.959711309 +0000 UTC m=+964.252998471" lastFinishedPulling="2026-03-21 00:27:51.588967075 +0000 UTC m=+984.882254247" observedRunningTime="2026-03-21 00:27:52.357317175 +0000 UTC m=+985.650604367" watchObservedRunningTime="2026-03-21 00:27:52.366678705 +0000 UTC m=+985.659965877" Mar 21 00:27:52 crc kubenswrapper[5117]: I0321 00:27:52.398142 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" podStartSLOduration=4.04705448 podStartE2EDuration="16.398126617s" podCreationTimestamp="2026-03-21 00:27:36 +0000 UTC" firstStartedPulling="2026-03-21 00:27:39.176734987 +0000 UTC m=+972.470022159" lastFinishedPulling="2026-03-21 00:27:51.527807124 +0000 UTC m=+984.821094296" observedRunningTime="2026-03-21 00:27:52.391350827 +0000 UTC m=+985.684638009" watchObservedRunningTime="2026-03-21 00:27:52.398126617 +0000 UTC m=+985.691413789" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.228749 5117 generic.go:358] "Generic (PLEG): container finished" podID="b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb" containerID="30f23160495446af3551d7451c46a512eeb7b95c5513bab2c653bf8272da4bda" exitCode=0 Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.228820 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" event={"ID":"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb","Type":"ContainerDied","Data":"30f23160495446af3551d7451c46a512eeb7b95c5513bab2c653bf8272da4bda"} Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.229093 5117 scope.go:117] "RemoveContainer" containerID="5ef900960997b2d513ec5718872126bd8e3f6b7f92518666bd8f01756f52cd50" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.229231 5117 scope.go:117] "RemoveContainer" containerID="30f23160495446af3551d7451c46a512eeb7b95c5513bab2c653bf8272da4bda" Mar 21 00:27:53 crc kubenswrapper[5117]: E0321 00:27:53.229988 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-6f48c45984-chc9c_service-telemetry(b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" podUID="b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.236834 5117 generic.go:358] "Generic (PLEG): container finished" podID="cd862294-3d4d-42a8-ba79-fc7607eac35e" containerID="5ea72378e01ebc98f30b831b529fa2af979904fca63c30b0286543a641d41733" exitCode=0 Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.236946 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" event={"ID":"cd862294-3d4d-42a8-ba79-fc7607eac35e","Type":"ContainerDied","Data":"5ea72378e01ebc98f30b831b529fa2af979904fca63c30b0286543a641d41733"} Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.237449 5117 scope.go:117] "RemoveContainer" containerID="5ea72378e01ebc98f30b831b529fa2af979904fca63c30b0286543a641d41733" Mar 21 00:27:53 crc kubenswrapper[5117]: E0321 00:27:53.237702 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-787645d794-qvvvj_service-telemetry(cd862294-3d4d-42a8-ba79-fc7607eac35e)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" podUID="cd862294-3d4d-42a8-ba79-fc7607eac35e" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.239772 5117 generic.go:358] "Generic (PLEG): container finished" podID="e6c95eaa-fc0b-48f0-93bf-1d2deefc2362" containerID="44723ea181acb256f9a1045f7c471a47bcb50f9a4e68425eb2986a794f99582e" exitCode=0 Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.239838 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" event={"ID":"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362","Type":"ContainerDied","Data":"44723ea181acb256f9a1045f7c471a47bcb50f9a4e68425eb2986a794f99582e"} Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.240093 5117 scope.go:117] "RemoveContainer" containerID="44723ea181acb256f9a1045f7c471a47bcb50f9a4e68425eb2986a794f99582e" Mar 21 00:27:53 crc kubenswrapper[5117]: E0321 00:27:53.240232 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289_service-telemetry(e6c95eaa-fc0b-48f0-93bf-1d2deefc2362)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" podUID="e6c95eaa-fc0b-48f0-93bf-1d2deefc2362" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.242163 5117 generic.go:358] "Generic (PLEG): container finished" podID="e1e39c35-dd28-4061-b744-debef1ae6e42" containerID="51ab96e2a99bbc9d35a9307ae47ab458e88a0f34986362e8afbfc113a3889a8b" exitCode=0 Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.242357 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" event={"ID":"e1e39c35-dd28-4061-b744-debef1ae6e42","Type":"ContainerDied","Data":"51ab96e2a99bbc9d35a9307ae47ab458e88a0f34986362e8afbfc113a3889a8b"} Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.242795 5117 scope.go:117] "RemoveContainer" containerID="51ab96e2a99bbc9d35a9307ae47ab458e88a0f34986362e8afbfc113a3889a8b" Mar 21 00:27:53 crc kubenswrapper[5117]: E0321 00:27:53.243042 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6_service-telemetry(e1e39c35-dd28-4061-b744-debef1ae6e42)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" podUID="e1e39c35-dd28-4061-b744-debef1ae6e42" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.245328 5117 generic.go:358] "Generic (PLEG): container finished" podID="cafef1e9-9a8d-4e86-b8ac-8480b2f141b4" containerID="256932e7300d587c1aab74c746b27af4ae2d4e1ca07e8defbe576b4d3c988349" exitCode=0 Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.246283 5117 scope.go:117] "RemoveContainer" containerID="256932e7300d587c1aab74c746b27af4ae2d4e1ca07e8defbe576b4d3c988349" Mar 21 00:27:53 crc kubenswrapper[5117]: E0321 00:27:53.246485 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth_service-telemetry(cafef1e9-9a8d-4e86-b8ac-8480b2f141b4)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" podUID="cafef1e9-9a8d-4e86-b8ac-8480b2f141b4" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.246527 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" event={"ID":"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4","Type":"ContainerDied","Data":"256932e7300d587c1aab74c746b27af4ae2d4e1ca07e8defbe576b4d3c988349"} Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.305146 5117 scope.go:117] "RemoveContainer" containerID="d9853fe504cf254d5533cdd3bbf732817f175fe6ee0783c4e75808f93676231b" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.375935 5117 scope.go:117] "RemoveContainer" containerID="f13f8c346488a4dc3ee3d945c33a582b1891e2d56b654c0e3d56f98f2eae6770" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.415902 5117 scope.go:117] "RemoveContainer" containerID="e03f1f1ef3145c8c131d7c8206dbdd1984572262516743ed38b5f8fb6a584e0d" Mar 21 00:27:53 crc kubenswrapper[5117]: I0321 00:27:53.451237 5117 scope.go:117] "RemoveContainer" containerID="404cb8b77731126a7b75429f875177e1803b8d6d318b220aed5b23959261ef59" Mar 21 00:27:54 crc kubenswrapper[5117]: I0321 00:27:54.256427 5117 scope.go:117] "RemoveContainer" containerID="30f23160495446af3551d7451c46a512eeb7b95c5513bab2c653bf8272da4bda" Mar 21 00:27:54 crc kubenswrapper[5117]: E0321 00:27:54.256736 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-6f48c45984-chc9c_service-telemetry(b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" podUID="b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb" Mar 21 00:27:54 crc kubenswrapper[5117]: I0321 00:27:54.258903 5117 scope.go:117] "RemoveContainer" containerID="5ea72378e01ebc98f30b831b529fa2af979904fca63c30b0286543a641d41733" Mar 21 00:27:54 crc kubenswrapper[5117]: E0321 00:27:54.259106 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-787645d794-qvvvj_service-telemetry(cd862294-3d4d-42a8-ba79-fc7607eac35e)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" podUID="cd862294-3d4d-42a8-ba79-fc7607eac35e" Mar 21 00:27:54 crc kubenswrapper[5117]: I0321 00:27:54.262108 5117 scope.go:117] "RemoveContainer" containerID="44723ea181acb256f9a1045f7c471a47bcb50f9a4e68425eb2986a794f99582e" Mar 21 00:27:54 crc kubenswrapper[5117]: E0321 00:27:54.262421 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289_service-telemetry(e6c95eaa-fc0b-48f0-93bf-1d2deefc2362)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" podUID="e6c95eaa-fc0b-48f0-93bf-1d2deefc2362" Mar 21 00:27:54 crc kubenswrapper[5117]: I0321 00:27:54.263969 5117 scope.go:117] "RemoveContainer" containerID="51ab96e2a99bbc9d35a9307ae47ab458e88a0f34986362e8afbfc113a3889a8b" Mar 21 00:27:54 crc kubenswrapper[5117]: E0321 00:27:54.264158 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6_service-telemetry(e1e39c35-dd28-4061-b744-debef1ae6e42)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" podUID="e1e39c35-dd28-4061-b744-debef1ae6e42" Mar 21 00:27:54 crc kubenswrapper[5117]: I0321 00:27:54.266992 5117 scope.go:117] "RemoveContainer" containerID="256932e7300d587c1aab74c746b27af4ae2d4e1ca07e8defbe576b4d3c988349" Mar 21 00:27:54 crc kubenswrapper[5117]: E0321 00:27:54.267184 5117 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth_service-telemetry(cafef1e9-9a8d-4e86-b8ac-8480b2f141b4)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" podUID="cafef1e9-9a8d-4e86-b8ac-8480b2f141b4" Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.133096 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567548-gfvwn"] Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.139028 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567548-gfvwn" Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.141204 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.141503 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.145658 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.147241 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567548-gfvwn"] Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.250039 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh4ct\" (UniqueName: \"kubernetes.io/projected/8b31000c-6a98-4e25-95cd-510fabb91ae2-kube-api-access-rh4ct\") pod \"auto-csr-approver-29567548-gfvwn\" (UID: \"8b31000c-6a98-4e25-95cd-510fabb91ae2\") " pod="openshift-infra/auto-csr-approver-29567548-gfvwn" Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.351749 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rh4ct\" (UniqueName: \"kubernetes.io/projected/8b31000c-6a98-4e25-95cd-510fabb91ae2-kube-api-access-rh4ct\") pod \"auto-csr-approver-29567548-gfvwn\" (UID: \"8b31000c-6a98-4e25-95cd-510fabb91ae2\") " pod="openshift-infra/auto-csr-approver-29567548-gfvwn" Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.370674 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh4ct\" (UniqueName: \"kubernetes.io/projected/8b31000c-6a98-4e25-95cd-510fabb91ae2-kube-api-access-rh4ct\") pod \"auto-csr-approver-29567548-gfvwn\" (UID: \"8b31000c-6a98-4e25-95cd-510fabb91ae2\") " pod="openshift-infra/auto-csr-approver-29567548-gfvwn" Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.455400 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567548-gfvwn" Mar 21 00:28:00 crc kubenswrapper[5117]: I0321 00:28:00.907969 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567548-gfvwn"] Mar 21 00:28:01 crc kubenswrapper[5117]: I0321 00:28:01.319986 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567548-gfvwn" event={"ID":"8b31000c-6a98-4e25-95cd-510fabb91ae2","Type":"ContainerStarted","Data":"d28b591b068e64409932a17f26926bf8260e186faaef9ef6bd53e6424242ea23"} Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.329689 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567548-gfvwn" event={"ID":"8b31000c-6a98-4e25-95cd-510fabb91ae2","Type":"ContainerStarted","Data":"15c5368178c60d1eff91c02a3a2cf4bee8a77669b672a0d1b26ac19f961204e8"} Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.344508 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567548-gfvwn" podStartSLOduration=1.372828895 podStartE2EDuration="2.344489742s" podCreationTimestamp="2026-03-21 00:28:00 +0000 UTC" firstStartedPulling="2026-03-21 00:28:00.92864638 +0000 UTC m=+994.221933542" lastFinishedPulling="2026-03-21 00:28:01.900307217 +0000 UTC m=+995.193594389" observedRunningTime="2026-03-21 00:28:02.341419247 +0000 UTC m=+995.634706449" watchObservedRunningTime="2026-03-21 00:28:02.344489742 +0000 UTC m=+995.637776934" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.443056 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.454337 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.454527 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.458785 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"qdr-test-config\"" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.458958 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-selfsigned\"" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.488177 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/0792ee3e-0a84-4937-a7ee-2b87dd42ec54-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"0792ee3e-0a84-4937-a7ee-2b87dd42ec54\") " pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.488258 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdt6l\" (UniqueName: \"kubernetes.io/projected/0792ee3e-0a84-4937-a7ee-2b87dd42ec54-kube-api-access-pdt6l\") pod \"qdr-test\" (UID: \"0792ee3e-0a84-4937-a7ee-2b87dd42ec54\") " pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.488327 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/0792ee3e-0a84-4937-a7ee-2b87dd42ec54-qdr-test-config\") pod \"qdr-test\" (UID: \"0792ee3e-0a84-4937-a7ee-2b87dd42ec54\") " pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.589874 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/0792ee3e-0a84-4937-a7ee-2b87dd42ec54-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"0792ee3e-0a84-4937-a7ee-2b87dd42ec54\") " pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.589975 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pdt6l\" (UniqueName: \"kubernetes.io/projected/0792ee3e-0a84-4937-a7ee-2b87dd42ec54-kube-api-access-pdt6l\") pod \"qdr-test\" (UID: \"0792ee3e-0a84-4937-a7ee-2b87dd42ec54\") " pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.590063 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/0792ee3e-0a84-4937-a7ee-2b87dd42ec54-qdr-test-config\") pod \"qdr-test\" (UID: \"0792ee3e-0a84-4937-a7ee-2b87dd42ec54\") " pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.591116 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/0792ee3e-0a84-4937-a7ee-2b87dd42ec54-qdr-test-config\") pod \"qdr-test\" (UID: \"0792ee3e-0a84-4937-a7ee-2b87dd42ec54\") " pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.595873 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/0792ee3e-0a84-4937-a7ee-2b87dd42ec54-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"0792ee3e-0a84-4937-a7ee-2b87dd42ec54\") " pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.612840 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdt6l\" (UniqueName: \"kubernetes.io/projected/0792ee3e-0a84-4937-a7ee-2b87dd42ec54-kube-api-access-pdt6l\") pod \"qdr-test\" (UID: \"0792ee3e-0a84-4937-a7ee-2b87dd42ec54\") " pod="service-telemetry/qdr-test" Mar 21 00:28:02 crc kubenswrapper[5117]: I0321 00:28:02.773562 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 21 00:28:03 crc kubenswrapper[5117]: I0321 00:28:03.180801 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 21 00:28:03 crc kubenswrapper[5117]: I0321 00:28:03.337738 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"0792ee3e-0a84-4937-a7ee-2b87dd42ec54","Type":"ContainerStarted","Data":"0cd1bfbbbd504a076d9f122a38365199af7b3adbee243a6f6fbdc1500fa08488"} Mar 21 00:28:03 crc kubenswrapper[5117]: I0321 00:28:03.339012 5117 generic.go:358] "Generic (PLEG): container finished" podID="8b31000c-6a98-4e25-95cd-510fabb91ae2" containerID="15c5368178c60d1eff91c02a3a2cf4bee8a77669b672a0d1b26ac19f961204e8" exitCode=0 Mar 21 00:28:03 crc kubenswrapper[5117]: I0321 00:28:03.339072 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567548-gfvwn" event={"ID":"8b31000c-6a98-4e25-95cd-510fabb91ae2","Type":"ContainerDied","Data":"15c5368178c60d1eff91c02a3a2cf4bee8a77669b672a0d1b26ac19f961204e8"} Mar 21 00:28:04 crc kubenswrapper[5117]: I0321 00:28:04.589868 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567548-gfvwn" Mar 21 00:28:04 crc kubenswrapper[5117]: I0321 00:28:04.621050 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh4ct\" (UniqueName: \"kubernetes.io/projected/8b31000c-6a98-4e25-95cd-510fabb91ae2-kube-api-access-rh4ct\") pod \"8b31000c-6a98-4e25-95cd-510fabb91ae2\" (UID: \"8b31000c-6a98-4e25-95cd-510fabb91ae2\") " Mar 21 00:28:04 crc kubenswrapper[5117]: I0321 00:28:04.645970 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b31000c-6a98-4e25-95cd-510fabb91ae2-kube-api-access-rh4ct" (OuterVolumeSpecName: "kube-api-access-rh4ct") pod "8b31000c-6a98-4e25-95cd-510fabb91ae2" (UID: "8b31000c-6a98-4e25-95cd-510fabb91ae2"). InnerVolumeSpecName "kube-api-access-rh4ct". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:28:04 crc kubenswrapper[5117]: I0321 00:28:04.722348 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rh4ct\" (UniqueName: \"kubernetes.io/projected/8b31000c-6a98-4e25-95cd-510fabb91ae2-kube-api-access-rh4ct\") on node \"crc\" DevicePath \"\"" Mar 21 00:28:05 crc kubenswrapper[5117]: I0321 00:28:05.357848 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567548-gfvwn" event={"ID":"8b31000c-6a98-4e25-95cd-510fabb91ae2","Type":"ContainerDied","Data":"d28b591b068e64409932a17f26926bf8260e186faaef9ef6bd53e6424242ea23"} Mar 21 00:28:05 crc kubenswrapper[5117]: I0321 00:28:05.357889 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d28b591b068e64409932a17f26926bf8260e186faaef9ef6bd53e6424242ea23" Mar 21 00:28:05 crc kubenswrapper[5117]: I0321 00:28:05.357894 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567548-gfvwn" Mar 21 00:28:05 crc kubenswrapper[5117]: I0321 00:28:05.417396 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567542-2fs89"] Mar 21 00:28:05 crc kubenswrapper[5117]: I0321 00:28:05.422503 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567542-2fs89"] Mar 21 00:28:05 crc kubenswrapper[5117]: I0321 00:28:05.591772 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff1f8f36-82ad-4771-b8de-b5dacf30a13a" path="/var/lib/kubelet/pods/ff1f8f36-82ad-4771-b8de-b5dacf30a13a/volumes" Mar 21 00:28:06 crc kubenswrapper[5117]: I0321 00:28:06.579836 5117 scope.go:117] "RemoveContainer" containerID="44723ea181acb256f9a1045f7c471a47bcb50f9a4e68425eb2986a794f99582e" Mar 21 00:28:07 crc kubenswrapper[5117]: I0321 00:28:07.586020 5117 scope.go:117] "RemoveContainer" containerID="256932e7300d587c1aab74c746b27af4ae2d4e1ca07e8defbe576b4d3c988349" Mar 21 00:28:08 crc kubenswrapper[5117]: I0321 00:28:08.578618 5117 scope.go:117] "RemoveContainer" containerID="51ab96e2a99bbc9d35a9307ae47ab458e88a0f34986362e8afbfc113a3889a8b" Mar 21 00:28:08 crc kubenswrapper[5117]: I0321 00:28:08.579983 5117 scope.go:117] "RemoveContainer" containerID="30f23160495446af3551d7451c46a512eeb7b95c5513bab2c653bf8272da4bda" Mar 21 00:28:09 crc kubenswrapper[5117]: I0321 00:28:09.578776 5117 scope.go:117] "RemoveContainer" containerID="5ea72378e01ebc98f30b831b529fa2af979904fca63c30b0286543a641d41733" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.443709 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f48c45984-chc9c" event={"ID":"b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb","Type":"ContainerStarted","Data":"b1e189a69787b42791f46edf59955682480386019eaed131e6aaebf8d1693c25"} Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.450369 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-qvvvj" event={"ID":"cd862294-3d4d-42a8-ba79-fc7607eac35e","Type":"ContainerStarted","Data":"53d4d3a256ea28fa8ae89ed331387db2171d8f05cb4f39a9b0bf9aff9377af4e"} Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.454701 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289" event={"ID":"e6c95eaa-fc0b-48f0-93bf-1d2deefc2362","Type":"ContainerStarted","Data":"293df4dc232412991937e1c734529a0e82ba6815617e604d207e5bbcf70d4f3c"} Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.459113 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6" event={"ID":"e1e39c35-dd28-4061-b744-debef1ae6e42","Type":"ContainerStarted","Data":"c0f7d881da0a9322908fb5bb6d36b7a5def977ba2e15c27f6e1052a016658a27"} Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.469800 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"0792ee3e-0a84-4937-a7ee-2b87dd42ec54","Type":"ContainerStarted","Data":"a69a64ddb6fc0a521d92a72a77d42be60bb714782e7613adcbda04c91ba79af0"} Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.475750 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth" event={"ID":"cafef1e9-9a8d-4e86-b8ac-8480b2f141b4","Type":"ContainerStarted","Data":"34169a1b2aeb71f8fc42cf420f7e2eff3a5c9f836e8bd5548a2f9ff74bfd72d8"} Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.583090 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=2.563842562 podStartE2EDuration="10.583063143s" podCreationTimestamp="2026-03-21 00:28:02 +0000 UTC" firstStartedPulling="2026-03-21 00:28:03.188086859 +0000 UTC m=+996.481374051" lastFinishedPulling="2026-03-21 00:28:11.20730746 +0000 UTC m=+1004.500594632" observedRunningTime="2026-03-21 00:28:12.577752949 +0000 UTC m=+1005.871040131" watchObservedRunningTime="2026-03-21 00:28:12.583063143 +0000 UTC m=+1005.876350315" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.845452 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-7sr6c"] Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.846590 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8b31000c-6a98-4e25-95cd-510fabb91ae2" containerName="oc" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.846621 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b31000c-6a98-4e25-95cd-510fabb91ae2" containerName="oc" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.846791 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="8b31000c-6a98-4e25-95cd-510fabb91ae2" containerName="oc" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.856885 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.859149 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-sensubility-config\"" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.859427 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-ceilometer-entrypoint-script\"" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.859474 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-collectd-config\"" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.859846 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-collectd-entrypoint-script\"" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.860554 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-healthcheck-log\"" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.860805 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-ceilometer-publisher\"" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.862736 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-7sr6c"] Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.956132 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-sensubility-config\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.956199 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-publisher\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.956265 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-config\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.956299 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.956354 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.956441 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2szfd\" (UniqueName: \"kubernetes.io/projected/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-kube-api-access-2szfd\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:12 crc kubenswrapper[5117]: I0321 00:28:12.956537 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-healthcheck-log\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.057469 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-config\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.057814 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.057848 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.057907 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2szfd\" (UniqueName: \"kubernetes.io/projected/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-kube-api-access-2szfd\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.057954 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-healthcheck-log\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.057976 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-sensubility-config\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.058002 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-publisher\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.058941 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-config\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.058949 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-publisher\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.059619 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.060108 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.060227 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-healthcheck-log\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.060825 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-sensubility-config\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.087022 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2szfd\" (UniqueName: \"kubernetes.io/projected/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-kube-api-access-2szfd\") pod \"stf-smoketest-smoke1-7sr6c\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.186819 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.211261 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.265494 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.265631 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.362642 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzmd2\" (UniqueName: \"kubernetes.io/projected/e1aba586-0bf3-4844-9358-b4ba919c2c36-kube-api-access-bzmd2\") pod \"curl\" (UID: \"e1aba586-0bf3-4844-9358-b4ba919c2c36\") " pod="service-telemetry/curl" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.463586 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bzmd2\" (UniqueName: \"kubernetes.io/projected/e1aba586-0bf3-4844-9358-b4ba919c2c36-kube-api-access-bzmd2\") pod \"curl\" (UID: \"e1aba586-0bf3-4844-9358-b4ba919c2c36\") " pod="service-telemetry/curl" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.482328 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzmd2\" (UniqueName: \"kubernetes.io/projected/e1aba586-0bf3-4844-9358-b4ba919c2c36-kube-api-access-bzmd2\") pod \"curl\" (UID: \"e1aba586-0bf3-4844-9358-b4ba919c2c36\") " pod="service-telemetry/curl" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.623896 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.711674 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-7sr6c"] Mar 21 00:28:13 crc kubenswrapper[5117]: W0321 00:28:13.726788 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a31ab38_af4e_4e1b_9c5c_e0ef429cdbca.slice/crio-ca317a7cf5bd4a11392455ab7eace56ad335faf99015f5ddd4a7887e63b264dd WatchSource:0}: Error finding container ca317a7cf5bd4a11392455ab7eace56ad335faf99015f5ddd4a7887e63b264dd: Status 404 returned error can't find the container with id ca317a7cf5bd4a11392455ab7eace56ad335faf99015f5ddd4a7887e63b264dd Mar 21 00:28:13 crc kubenswrapper[5117]: I0321 00:28:13.867739 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 21 00:28:13 crc kubenswrapper[5117]: W0321 00:28:13.870389 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1aba586_0bf3_4844_9358_b4ba919c2c36.slice/crio-8103dacc6322083929209dc045d082d39583b5fea61f3148cab8cdf8c814da08 WatchSource:0}: Error finding container 8103dacc6322083929209dc045d082d39583b5fea61f3148cab8cdf8c814da08: Status 404 returned error can't find the container with id 8103dacc6322083929209dc045d082d39583b5fea61f3148cab8cdf8c814da08 Mar 21 00:28:14 crc kubenswrapper[5117]: I0321 00:28:14.495692 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"e1aba586-0bf3-4844-9358-b4ba919c2c36","Type":"ContainerStarted","Data":"8103dacc6322083929209dc045d082d39583b5fea61f3148cab8cdf8c814da08"} Mar 21 00:28:14 crc kubenswrapper[5117]: I0321 00:28:14.497164 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" event={"ID":"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca","Type":"ContainerStarted","Data":"ca317a7cf5bd4a11392455ab7eace56ad335faf99015f5ddd4a7887e63b264dd"} Mar 21 00:28:16 crc kubenswrapper[5117]: I0321 00:28:16.516179 5117 generic.go:358] "Generic (PLEG): container finished" podID="e1aba586-0bf3-4844-9358-b4ba919c2c36" containerID="3188850db87fe971d60b3ea1ddb431c3be83fb6fe982de2dc930fb9f9be987d7" exitCode=0 Mar 21 00:28:16 crc kubenswrapper[5117]: I0321 00:28:16.516220 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"e1aba586-0bf3-4844-9358-b4ba919c2c36","Type":"ContainerDied","Data":"3188850db87fe971d60b3ea1ddb431c3be83fb6fe982de2dc930fb9f9be987d7"} Mar 21 00:28:20 crc kubenswrapper[5117]: I0321 00:28:20.218394 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 21 00:28:20 crc kubenswrapper[5117]: I0321 00:28:20.373967 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzmd2\" (UniqueName: \"kubernetes.io/projected/e1aba586-0bf3-4844-9358-b4ba919c2c36-kube-api-access-bzmd2\") pod \"e1aba586-0bf3-4844-9358-b4ba919c2c36\" (UID: \"e1aba586-0bf3-4844-9358-b4ba919c2c36\") " Mar 21 00:28:20 crc kubenswrapper[5117]: I0321 00:28:20.381942 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1aba586-0bf3-4844-9358-b4ba919c2c36-kube-api-access-bzmd2" (OuterVolumeSpecName: "kube-api-access-bzmd2") pod "e1aba586-0bf3-4844-9358-b4ba919c2c36" (UID: "e1aba586-0bf3-4844-9358-b4ba919c2c36"). InnerVolumeSpecName "kube-api-access-bzmd2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:28:20 crc kubenswrapper[5117]: I0321 00:28:20.406666 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_e1aba586-0bf3-4844-9358-b4ba919c2c36/curl/0.log" Mar 21 00:28:20 crc kubenswrapper[5117]: I0321 00:28:20.476448 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-bzmd2\" (UniqueName: \"kubernetes.io/projected/e1aba586-0bf3-4844-9358-b4ba919c2c36-kube-api-access-bzmd2\") on node \"crc\" DevicePath \"\"" Mar 21 00:28:20 crc kubenswrapper[5117]: I0321 00:28:20.549749 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"e1aba586-0bf3-4844-9358-b4ba919c2c36","Type":"ContainerDied","Data":"8103dacc6322083929209dc045d082d39583b5fea61f3148cab8cdf8c814da08"} Mar 21 00:28:20 crc kubenswrapper[5117]: I0321 00:28:20.549790 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8103dacc6322083929209dc045d082d39583b5fea61f3148cab8cdf8c814da08" Mar 21 00:28:20 crc kubenswrapper[5117]: I0321 00:28:20.549796 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 21 00:28:20 crc kubenswrapper[5117]: I0321 00:28:20.668989 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6774d8dfbc-9mdbs_b4dbbbf2-a573-41be-b77c-4beba9161a98/prometheus-webhook-snmp/0.log" Mar 21 00:28:25 crc kubenswrapper[5117]: I0321 00:28:25.593447 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" event={"ID":"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca","Type":"ContainerStarted","Data":"8bb10eb664961e3fabbb2ab81c19ee2cb2dd79a1f187c6f89f69b069accf652e"} Mar 21 00:28:30 crc kubenswrapper[5117]: I0321 00:28:30.146919 5117 scope.go:117] "RemoveContainer" containerID="42dfd97a193e2da73a25b9bd568728ae5b1e21c466dee8e8bb596985024ae4ec" Mar 21 00:28:31 crc kubenswrapper[5117]: I0321 00:28:31.649432 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" event={"ID":"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca","Type":"ContainerStarted","Data":"821b58ca50830c8dee03ea1218293fffe99e9c0f46e14a8d4280f2c5b2f38fec"} Mar 21 00:28:31 crc kubenswrapper[5117]: I0321 00:28:31.673789 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" podStartSLOduration=2.15940261 podStartE2EDuration="19.673768413s" podCreationTimestamp="2026-03-21 00:28:12 +0000 UTC" firstStartedPulling="2026-03-21 00:28:13.728831975 +0000 UTC m=+1007.022119157" lastFinishedPulling="2026-03-21 00:28:31.243197788 +0000 UTC m=+1024.536484960" observedRunningTime="2026-03-21 00:28:31.668442388 +0000 UTC m=+1024.961729580" watchObservedRunningTime="2026-03-21 00:28:31.673768413 +0000 UTC m=+1024.967055585" Mar 21 00:28:50 crc kubenswrapper[5117]: I0321 00:28:50.861729 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6774d8dfbc-9mdbs_b4dbbbf2-a573-41be-b77c-4beba9161a98/prometheus-webhook-snmp/0.log" Mar 21 00:28:58 crc kubenswrapper[5117]: I0321 00:28:58.930435 5117 generic.go:358] "Generic (PLEG): container finished" podID="1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" containerID="8bb10eb664961e3fabbb2ab81c19ee2cb2dd79a1f187c6f89f69b069accf652e" exitCode=0 Mar 21 00:28:58 crc kubenswrapper[5117]: I0321 00:28:58.931326 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" event={"ID":"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca","Type":"ContainerDied","Data":"8bb10eb664961e3fabbb2ab81c19ee2cb2dd79a1f187c6f89f69b069accf652e"} Mar 21 00:28:58 crc kubenswrapper[5117]: I0321 00:28:58.931986 5117 scope.go:117] "RemoveContainer" containerID="8bb10eb664961e3fabbb2ab81c19ee2cb2dd79a1f187c6f89f69b069accf652e" Mar 21 00:29:03 crc kubenswrapper[5117]: I0321 00:29:03.970601 5117 generic.go:358] "Generic (PLEG): container finished" podID="1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" containerID="821b58ca50830c8dee03ea1218293fffe99e9c0f46e14a8d4280f2c5b2f38fec" exitCode=0 Mar 21 00:29:03 crc kubenswrapper[5117]: I0321 00:29:03.970908 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" event={"ID":"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca","Type":"ContainerDied","Data":"821b58ca50830c8dee03ea1218293fffe99e9c0f46e14a8d4280f2c5b2f38fec"} Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.330807 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.431516 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-config\") pod \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.431583 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-entrypoint-script\") pod \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.431655 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-entrypoint-script\") pod \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.431712 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-healthcheck-log\") pod \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.431753 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-publisher\") pod \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.431819 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2szfd\" (UniqueName: \"kubernetes.io/projected/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-kube-api-access-2szfd\") pod \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.431888 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-sensubility-config\") pod \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\" (UID: \"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca\") " Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.444499 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-kube-api-access-2szfd" (OuterVolumeSpecName: "kube-api-access-2szfd") pod "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" (UID: "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca"). InnerVolumeSpecName "kube-api-access-2szfd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.452477 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" (UID: "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.454024 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" (UID: "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.456311 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" (UID: "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.456981 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" (UID: "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.465423 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" (UID: "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.466604 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" (UID: "1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.533571 5117 reconciler_common.go:299] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-healthcheck-log\") on node \"crc\" DevicePath \"\"" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.533630 5117 reconciler_common.go:299] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.533655 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2szfd\" (UniqueName: \"kubernetes.io/projected/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-kube-api-access-2szfd\") on node \"crc\" DevicePath \"\"" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.533675 5117 reconciler_common.go:299] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-sensubility-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.533695 5117 reconciler_common.go:299] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-config\") on node \"crc\" DevicePath \"\"" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.533712 5117 reconciler_common.go:299] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.533730 5117 reconciler_common.go:299] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.990007 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" event={"ID":"1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca","Type":"ContainerDied","Data":"ca317a7cf5bd4a11392455ab7eace56ad335faf99015f5ddd4a7887e63b264dd"} Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.990282 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca317a7cf5bd4a11392455ab7eace56ad335faf99015f5ddd4a7887e63b264dd" Mar 21 00:29:05 crc kubenswrapper[5117]: I0321 00:29:05.990425 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-7sr6c" Mar 21 00:29:07 crc kubenswrapper[5117]: I0321 00:29:07.395450 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-7sr6c_1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca/smoketest-collectd/0.log" Mar 21 00:29:07 crc kubenswrapper[5117]: I0321 00:29:07.718203 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-7sr6c_1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca/smoketest-ceilometer/0.log" Mar 21 00:29:08 crc kubenswrapper[5117]: I0321 00:29:08.018740 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-55bf8d5cb-x4qfv_b62f948f-1381-4e90-a11c-e5ec089639ef/default-interconnect/0.log" Mar 21 00:29:08 crc kubenswrapper[5117]: I0321 00:29:08.375993 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-787645d794-qvvvj_cd862294-3d4d-42a8-ba79-fc7607eac35e/bridge/2.log" Mar 21 00:29:08 crc kubenswrapper[5117]: I0321 00:29:08.686888 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-787645d794-qvvvj_cd862294-3d4d-42a8-ba79-fc7607eac35e/sg-core/0.log" Mar 21 00:29:08 crc kubenswrapper[5117]: I0321 00:29:08.990578 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-6f48c45984-chc9c_b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb/bridge/2.log" Mar 21 00:29:09 crc kubenswrapper[5117]: I0321 00:29:09.274499 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-6f48c45984-chc9c_b1c9e2b8-36c4-495c-95d5-a04cbdd00bdb/sg-core/0.log" Mar 21 00:29:09 crc kubenswrapper[5117]: I0321 00:29:09.557040 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289_e6c95eaa-fc0b-48f0-93bf-1d2deefc2362/bridge/2.log" Mar 21 00:29:09 crc kubenswrapper[5117]: I0321 00:29:09.816411 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-545b564d9f-wg289_e6c95eaa-fc0b-48f0-93bf-1d2deefc2362/sg-core/0.log" Mar 21 00:29:10 crc kubenswrapper[5117]: I0321 00:29:10.139161 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6_e1e39c35-dd28-4061-b744-debef1ae6e42/bridge/2.log" Mar 21 00:29:10 crc kubenswrapper[5117]: I0321 00:29:10.422600 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-79976777f5-r9pd6_e1e39c35-dd28-4061-b744-debef1ae6e42/sg-core/0.log" Mar 21 00:29:10 crc kubenswrapper[5117]: I0321 00:29:10.717415 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth_cafef1e9-9a8d-4e86-b8ac-8480b2f141b4/bridge/2.log" Mar 21 00:29:11 crc kubenswrapper[5117]: I0321 00:29:11.026933 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-66d5b7c5fc-pxkth_cafef1e9-9a8d-4e86-b8ac-8480b2f141b4/sg-core/0.log" Mar 21 00:29:14 crc kubenswrapper[5117]: I0321 00:29:14.339045 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-79dd94fcb7-tkcvb_51dea3cf-4c01-4c26-99e7-d5bebf6589d4/operator/0.log" Mar 21 00:29:14 crc kubenswrapper[5117]: I0321 00:29:14.661766 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_b1f577e0-5a48-4c25-bbe2-7d20de72553f/prometheus/0.log" Mar 21 00:29:15 crc kubenswrapper[5117]: I0321 00:29:15.041443 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_ca7eaf9c-252d-4169-91c0-33d83751a9ae/elasticsearch/0.log" Mar 21 00:29:15 crc kubenswrapper[5117]: I0321 00:29:15.294030 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6774d8dfbc-9mdbs_b4dbbbf2-a573-41be-b77c-4beba9161a98/prometheus-webhook-snmp/0.log" Mar 21 00:29:15 crc kubenswrapper[5117]: I0321 00:29:15.628622 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_cd169b92-c7d5-4c3a-a6ce-db74aa6422ed/alertmanager/0.log" Mar 21 00:29:24 crc kubenswrapper[5117]: I0321 00:29:24.718583 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:29:24 crc kubenswrapper[5117]: I0321 00:29:24.719224 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:29:29 crc kubenswrapper[5117]: I0321 00:29:29.162339 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-b555588b5-tfz84_647ba439-f6c5-4a74-890f-28f776a83520/operator/0.log" Mar 21 00:29:32 crc kubenswrapper[5117]: I0321 00:29:32.875997 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-79dd94fcb7-tkcvb_51dea3cf-4c01-4c26-99e7-d5bebf6589d4/operator/0.log" Mar 21 00:29:33 crc kubenswrapper[5117]: I0321 00:29:33.164742 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_0792ee3e-0a84-4937-a7ee-2b87dd42ec54/qdr/0.log" Mar 21 00:29:54 crc kubenswrapper[5117]: I0321 00:29:54.718563 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:29:54 crc kubenswrapper[5117]: I0321 00:29:54.719329 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.152956 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567550-v4nrj"] Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.155247 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e1aba586-0bf3-4844-9358-b4ba919c2c36" containerName="curl" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.155334 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1aba586-0bf3-4844-9358-b4ba919c2c36" containerName="curl" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.155377 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" containerName="smoketest-ceilometer" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.155390 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" containerName="smoketest-ceilometer" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.155423 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" containerName="smoketest-collectd" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.155436 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" containerName="smoketest-collectd" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.155678 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="e1aba586-0bf3-4844-9358-b4ba919c2c36" containerName="curl" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.155702 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" containerName="smoketest-ceilometer" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.155720 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="1a31ab38-af4e-4e1b-9c5c-e0ef429cdbca" containerName="smoketest-collectd" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.163659 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm"] Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.164081 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567550-v4nrj" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.167632 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.167940 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.168297 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.172708 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567550-v4nrj"] Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.172960 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.175683 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm"] Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.180195 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-config\"" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.180311 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-dockercfg-vfqp6\"" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.242588 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce97026c-6936-4818-a521-444c3d0ea16a-secret-volume\") pod \"collect-profiles-29567550-fh6gm\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.242671 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksm9d\" (UniqueName: \"kubernetes.io/projected/9e18a4af-8e0f-450e-b461-90ba1217230e-kube-api-access-ksm9d\") pod \"auto-csr-approver-29567550-v4nrj\" (UID: \"9e18a4af-8e0f-450e-b461-90ba1217230e\") " pod="openshift-infra/auto-csr-approver-29567550-v4nrj" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.242706 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28kcv\" (UniqueName: \"kubernetes.io/projected/ce97026c-6936-4818-a521-444c3d0ea16a-kube-api-access-28kcv\") pod \"collect-profiles-29567550-fh6gm\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.242857 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce97026c-6936-4818-a521-444c3d0ea16a-config-volume\") pod \"collect-profiles-29567550-fh6gm\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.343645 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ksm9d\" (UniqueName: \"kubernetes.io/projected/9e18a4af-8e0f-450e-b461-90ba1217230e-kube-api-access-ksm9d\") pod \"auto-csr-approver-29567550-v4nrj\" (UID: \"9e18a4af-8e0f-450e-b461-90ba1217230e\") " pod="openshift-infra/auto-csr-approver-29567550-v4nrj" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.343710 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-28kcv\" (UniqueName: \"kubernetes.io/projected/ce97026c-6936-4818-a521-444c3d0ea16a-kube-api-access-28kcv\") pod \"collect-profiles-29567550-fh6gm\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.343819 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce97026c-6936-4818-a521-444c3d0ea16a-config-volume\") pod \"collect-profiles-29567550-fh6gm\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.343870 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce97026c-6936-4818-a521-444c3d0ea16a-secret-volume\") pod \"collect-profiles-29567550-fh6gm\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.345242 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce97026c-6936-4818-a521-444c3d0ea16a-config-volume\") pod \"collect-profiles-29567550-fh6gm\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.358152 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce97026c-6936-4818-a521-444c3d0ea16a-secret-volume\") pod \"collect-profiles-29567550-fh6gm\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.362189 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksm9d\" (UniqueName: \"kubernetes.io/projected/9e18a4af-8e0f-450e-b461-90ba1217230e-kube-api-access-ksm9d\") pod \"auto-csr-approver-29567550-v4nrj\" (UID: \"9e18a4af-8e0f-450e-b461-90ba1217230e\") " pod="openshift-infra/auto-csr-approver-29567550-v4nrj" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.385797 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-28kcv\" (UniqueName: \"kubernetes.io/projected/ce97026c-6936-4818-a521-444c3d0ea16a-kube-api-access-28kcv\") pod \"collect-profiles-29567550-fh6gm\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.488755 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567550-v4nrj" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.497382 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.779638 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567550-v4nrj"] Mar 21 00:30:00 crc kubenswrapper[5117]: I0321 00:30:00.820613 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm"] Mar 21 00:30:00 crc kubenswrapper[5117]: W0321 00:30:00.825194 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce97026c_6936_4818_a521_444c3d0ea16a.slice/crio-d724ce1707195a603719da95f84262131cea7ac00fb162824609c26914ca0ed3 WatchSource:0}: Error finding container d724ce1707195a603719da95f84262131cea7ac00fb162824609c26914ca0ed3: Status 404 returned error can't find the container with id d724ce1707195a603719da95f84262131cea7ac00fb162824609c26914ca0ed3 Mar 21 00:30:01 crc kubenswrapper[5117]: I0321 00:30:01.570243 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567550-v4nrj" event={"ID":"9e18a4af-8e0f-450e-b461-90ba1217230e","Type":"ContainerStarted","Data":"71b476a8408c2f8a7a7a708a750a364e9360aae366260b412af612872b111494"} Mar 21 00:30:01 crc kubenswrapper[5117]: I0321 00:30:01.573447 5117 generic.go:358] "Generic (PLEG): container finished" podID="ce97026c-6936-4818-a521-444c3d0ea16a" containerID="c29f9ee2e8cb50fb601336bf2ba0451110afbef6e9154f04941e55c3f739c127" exitCode=0 Mar 21 00:30:01 crc kubenswrapper[5117]: I0321 00:30:01.573770 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" event={"ID":"ce97026c-6936-4818-a521-444c3d0ea16a","Type":"ContainerDied","Data":"c29f9ee2e8cb50fb601336bf2ba0451110afbef6e9154f04941e55c3f739c127"} Mar 21 00:30:01 crc kubenswrapper[5117]: I0321 00:30:01.573834 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" event={"ID":"ce97026c-6936-4818-a521-444c3d0ea16a","Type":"ContainerStarted","Data":"d724ce1707195a603719da95f84262131cea7ac00fb162824609c26914ca0ed3"} Mar 21 00:30:02 crc kubenswrapper[5117]: I0321 00:30:02.583093 5117 generic.go:358] "Generic (PLEG): container finished" podID="9e18a4af-8e0f-450e-b461-90ba1217230e" containerID="8da50d572e4c5ee934f65636c7052b1607d34eb45159dba3019f27ff57c91d33" exitCode=0 Mar 21 00:30:02 crc kubenswrapper[5117]: I0321 00:30:02.583208 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567550-v4nrj" event={"ID":"9e18a4af-8e0f-450e-b461-90ba1217230e","Type":"ContainerDied","Data":"8da50d572e4c5ee934f65636c7052b1607d34eb45159dba3019f27ff57c91d33"} Mar 21 00:30:02 crc kubenswrapper[5117]: I0321 00:30:02.896490 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:02 crc kubenswrapper[5117]: I0321 00:30:02.987309 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce97026c-6936-4818-a521-444c3d0ea16a-secret-volume\") pod \"ce97026c-6936-4818-a521-444c3d0ea16a\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " Mar 21 00:30:02 crc kubenswrapper[5117]: I0321 00:30:02.987525 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28kcv\" (UniqueName: \"kubernetes.io/projected/ce97026c-6936-4818-a521-444c3d0ea16a-kube-api-access-28kcv\") pod \"ce97026c-6936-4818-a521-444c3d0ea16a\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " Mar 21 00:30:02 crc kubenswrapper[5117]: I0321 00:30:02.987640 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce97026c-6936-4818-a521-444c3d0ea16a-config-volume\") pod \"ce97026c-6936-4818-a521-444c3d0ea16a\" (UID: \"ce97026c-6936-4818-a521-444c3d0ea16a\") " Mar 21 00:30:02 crc kubenswrapper[5117]: I0321 00:30:02.988306 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce97026c-6936-4818-a521-444c3d0ea16a-config-volume" (OuterVolumeSpecName: "config-volume") pod "ce97026c-6936-4818-a521-444c3d0ea16a" (UID: "ce97026c-6936-4818-a521-444c3d0ea16a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 21 00:30:02 crc kubenswrapper[5117]: I0321 00:30:02.996965 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce97026c-6936-4818-a521-444c3d0ea16a-kube-api-access-28kcv" (OuterVolumeSpecName: "kube-api-access-28kcv") pod "ce97026c-6936-4818-a521-444c3d0ea16a" (UID: "ce97026c-6936-4818-a521-444c3d0ea16a"). InnerVolumeSpecName "kube-api-access-28kcv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:30:02 crc kubenswrapper[5117]: I0321 00:30:02.997142 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce97026c-6936-4818-a521-444c3d0ea16a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ce97026c-6936-4818-a521-444c3d0ea16a" (UID: "ce97026c-6936-4818-a521-444c3d0ea16a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 21 00:30:03 crc kubenswrapper[5117]: I0321 00:30:03.089858 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-28kcv\" (UniqueName: \"kubernetes.io/projected/ce97026c-6936-4818-a521-444c3d0ea16a-kube-api-access-28kcv\") on node \"crc\" DevicePath \"\"" Mar 21 00:30:03 crc kubenswrapper[5117]: I0321 00:30:03.089902 5117 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce97026c-6936-4818-a521-444c3d0ea16a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 21 00:30:03 crc kubenswrapper[5117]: I0321 00:30:03.089919 5117 reconciler_common.go:299] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce97026c-6936-4818-a521-444c3d0ea16a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 21 00:30:03 crc kubenswrapper[5117]: I0321 00:30:03.595799 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" Mar 21 00:30:03 crc kubenswrapper[5117]: I0321 00:30:03.595848 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29567550-fh6gm" event={"ID":"ce97026c-6936-4818-a521-444c3d0ea16a","Type":"ContainerDied","Data":"d724ce1707195a603719da95f84262131cea7ac00fb162824609c26914ca0ed3"} Mar 21 00:30:03 crc kubenswrapper[5117]: I0321 00:30:03.595885 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d724ce1707195a603719da95f84262131cea7ac00fb162824609c26914ca0ed3" Mar 21 00:30:03 crc kubenswrapper[5117]: E0321 00:30:03.794544 5117 cadvisor_stats_provider.go:525] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce97026c_6936_4818_a521_444c3d0ea16a.slice\": RecentStats: unable to find data in memory cache]" Mar 21 00:30:03 crc kubenswrapper[5117]: I0321 00:30:03.961533 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567550-v4nrj" Mar 21 00:30:04 crc kubenswrapper[5117]: I0321 00:30:04.109238 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksm9d\" (UniqueName: \"kubernetes.io/projected/9e18a4af-8e0f-450e-b461-90ba1217230e-kube-api-access-ksm9d\") pod \"9e18a4af-8e0f-450e-b461-90ba1217230e\" (UID: \"9e18a4af-8e0f-450e-b461-90ba1217230e\") " Mar 21 00:30:04 crc kubenswrapper[5117]: I0321 00:30:04.118590 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e18a4af-8e0f-450e-b461-90ba1217230e-kube-api-access-ksm9d" (OuterVolumeSpecName: "kube-api-access-ksm9d") pod "9e18a4af-8e0f-450e-b461-90ba1217230e" (UID: "9e18a4af-8e0f-450e-b461-90ba1217230e"). InnerVolumeSpecName "kube-api-access-ksm9d". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:30:04 crc kubenswrapper[5117]: I0321 00:30:04.212342 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ksm9d\" (UniqueName: \"kubernetes.io/projected/9e18a4af-8e0f-450e-b461-90ba1217230e-kube-api-access-ksm9d\") on node \"crc\" DevicePath \"\"" Mar 21 00:30:04 crc kubenswrapper[5117]: I0321 00:30:04.607218 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567550-v4nrj" Mar 21 00:30:04 crc kubenswrapper[5117]: I0321 00:30:04.607322 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567550-v4nrj" event={"ID":"9e18a4af-8e0f-450e-b461-90ba1217230e","Type":"ContainerDied","Data":"71b476a8408c2f8a7a7a708a750a364e9360aae366260b412af612872b111494"} Mar 21 00:30:04 crc kubenswrapper[5117]: I0321 00:30:04.607395 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71b476a8408c2f8a7a7a708a750a364e9360aae366260b412af612872b111494" Mar 21 00:30:05 crc kubenswrapper[5117]: I0321 00:30:05.032990 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567544-rmr2v"] Mar 21 00:30:05 crc kubenswrapper[5117]: I0321 00:30:05.039049 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567544-rmr2v"] Mar 21 00:30:05 crc kubenswrapper[5117]: I0321 00:30:05.591172 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be0be9be-183f-44ae-bea1-13b4c356bd8c" path="/var/lib/kubelet/pods/be0be9be-183f-44ae-bea1-13b4c356bd8c/volumes" Mar 21 00:30:08 crc kubenswrapper[5117]: I0321 00:30:08.986454 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jj5pz/must-gather-vfn22"] Mar 21 00:30:08 crc kubenswrapper[5117]: I0321 00:30:08.987621 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9e18a4af-8e0f-450e-b461-90ba1217230e" containerName="oc" Mar 21 00:30:08 crc kubenswrapper[5117]: I0321 00:30:08.987638 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e18a4af-8e0f-450e-b461-90ba1217230e" containerName="oc" Mar 21 00:30:08 crc kubenswrapper[5117]: I0321 00:30:08.987667 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ce97026c-6936-4818-a521-444c3d0ea16a" containerName="collect-profiles" Mar 21 00:30:08 crc kubenswrapper[5117]: I0321 00:30:08.987675 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce97026c-6936-4818-a521-444c3d0ea16a" containerName="collect-profiles" Mar 21 00:30:08 crc kubenswrapper[5117]: I0321 00:30:08.987846 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="ce97026c-6936-4818-a521-444c3d0ea16a" containerName="collect-profiles" Mar 21 00:30:08 crc kubenswrapper[5117]: I0321 00:30:08.987861 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="9e18a4af-8e0f-450e-b461-90ba1217230e" containerName="oc" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.000861 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.001457 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jj5pz/must-gather-vfn22"] Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.003786 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-jj5pz\"/\"openshift-service-ca.crt\"" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.004047 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-jj5pz\"/\"kube-root-ca.crt\"" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.004163 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-jj5pz\"/\"default-dockercfg-5vsvj\"" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.090605 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-must-gather-output\") pod \"must-gather-vfn22\" (UID: \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\") " pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.090936 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vngcs\" (UniqueName: \"kubernetes.io/projected/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-kube-api-access-vngcs\") pod \"must-gather-vfn22\" (UID: \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\") " pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.192558 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vngcs\" (UniqueName: \"kubernetes.io/projected/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-kube-api-access-vngcs\") pod \"must-gather-vfn22\" (UID: \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\") " pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.192733 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-must-gather-output\") pod \"must-gather-vfn22\" (UID: \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\") " pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.193131 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-must-gather-output\") pod \"must-gather-vfn22\" (UID: \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\") " pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.212175 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vngcs\" (UniqueName: \"kubernetes.io/projected/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-kube-api-access-vngcs\") pod \"must-gather-vfn22\" (UID: \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\") " pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.324545 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.542899 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jj5pz/must-gather-vfn22"] Mar 21 00:30:09 crc kubenswrapper[5117]: W0321 00:30:09.558853 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6afcbed0_7a3d_4a2a_be56_de45f60b34d4.slice/crio-db38b2702cdcd66638b544ddc9918b6f39c1eb4a2cea7f835306abc6f67867e7 WatchSource:0}: Error finding container db38b2702cdcd66638b544ddc9918b6f39c1eb4a2cea7f835306abc6f67867e7: Status 404 returned error can't find the container with id db38b2702cdcd66638b544ddc9918b6f39c1eb4a2cea7f835306abc6f67867e7 Mar 21 00:30:09 crc kubenswrapper[5117]: I0321 00:30:09.654479 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jj5pz/must-gather-vfn22" event={"ID":"6afcbed0-7a3d-4a2a-be56-de45f60b34d4","Type":"ContainerStarted","Data":"db38b2702cdcd66638b544ddc9918b6f39c1eb4a2cea7f835306abc6f67867e7"} Mar 21 00:30:16 crc kubenswrapper[5117]: I0321 00:30:16.719077 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jj5pz/must-gather-vfn22" event={"ID":"6afcbed0-7a3d-4a2a-be56-de45f60b34d4","Type":"ContainerStarted","Data":"5f46110e5afd0ae54d546a7780fc58d1f2fbca2a603e7d155817599c9af7c669"} Mar 21 00:30:16 crc kubenswrapper[5117]: I0321 00:30:16.719466 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jj5pz/must-gather-vfn22" event={"ID":"6afcbed0-7a3d-4a2a-be56-de45f60b34d4","Type":"ContainerStarted","Data":"5568cca41b73321c702ac54da1df361182483319e5c8df59523d75b1ce174c4f"} Mar 21 00:30:24 crc kubenswrapper[5117]: I0321 00:30:24.718912 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:30:24 crc kubenswrapper[5117]: I0321 00:30:24.719552 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:30:24 crc kubenswrapper[5117]: I0321 00:30:24.719609 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:30:24 crc kubenswrapper[5117]: I0321 00:30:24.720272 5117 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb586c1751d61bfe9aa828e1aeabf26cbd66967a6d89d22ac2c57498df26a314"} pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 00:30:24 crc kubenswrapper[5117]: I0321 00:30:24.720340 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" containerID="cri-o://eb586c1751d61bfe9aa828e1aeabf26cbd66967a6d89d22ac2c57498df26a314" gracePeriod=600 Mar 21 00:30:25 crc kubenswrapper[5117]: I0321 00:30:25.793040 5117 generic.go:358] "Generic (PLEG): container finished" podID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerID="eb586c1751d61bfe9aa828e1aeabf26cbd66967a6d89d22ac2c57498df26a314" exitCode=0 Mar 21 00:30:25 crc kubenswrapper[5117]: I0321 00:30:25.793132 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerDied","Data":"eb586c1751d61bfe9aa828e1aeabf26cbd66967a6d89d22ac2c57498df26a314"} Mar 21 00:30:25 crc kubenswrapper[5117]: I0321 00:30:25.793499 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerStarted","Data":"7bf5d1dd104623548e3c0f4587a917340dbfe836ba859055e5a1a9dbd9072797"} Mar 21 00:30:25 crc kubenswrapper[5117]: I0321 00:30:25.793522 5117 scope.go:117] "RemoveContainer" containerID="324331d7c0a640c4e69a7d08e2bb3a2ebe5299eeda584b902a33658b2293b76e" Mar 21 00:30:25 crc kubenswrapper[5117]: I0321 00:30:25.817423 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jj5pz/must-gather-vfn22" podStartSLOduration=11.825165517 podStartE2EDuration="17.81740535s" podCreationTimestamp="2026-03-21 00:30:08 +0000 UTC" firstStartedPulling="2026-03-21 00:30:09.560730385 +0000 UTC m=+1122.854017557" lastFinishedPulling="2026-03-21 00:30:15.552970178 +0000 UTC m=+1128.846257390" observedRunningTime="2026-03-21 00:30:16.746843815 +0000 UTC m=+1130.040130997" watchObservedRunningTime="2026-03-21 00:30:25.81740535 +0000 UTC m=+1139.110692522" Mar 21 00:30:31 crc kubenswrapper[5117]: I0321 00:30:31.024229 5117 scope.go:117] "RemoveContainer" containerID="49362447ce32884053fa951ea528635bce3198f9644e27f94d27b0227b67d580" Mar 21 00:31:06 crc kubenswrapper[5117]: I0321 00:31:06.193126 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-75ffdb6fcd-qt6hv_ee00846f-ea12-4777-a333-942c9e5e7cea/control-plane-machine-set-operator/0.log" Mar 21 00:31:06 crc kubenswrapper[5117]: I0321 00:31:06.332509 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-755bb95488-d9bzl_828ac2ae-7b43-4714-9cb9-a86252d5deeb/kube-rbac-proxy/0.log" Mar 21 00:31:06 crc kubenswrapper[5117]: I0321 00:31:06.377734 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-755bb95488-d9bzl_828ac2ae-7b43-4714-9cb9-a86252d5deeb/machine-api-operator/0.log" Mar 21 00:31:19 crc kubenswrapper[5117]: I0321 00:31:19.980084 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-759f64656b-9p6vr_27d8650b-ad7e-452a-b1d2-5605dcfe437a/cert-manager-controller/0.log" Mar 21 00:31:20 crc kubenswrapper[5117]: I0321 00:31:20.163235 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-8966b78d4-z7sqv_2aa9faf9-ec74-49e6-a0de-48c5e938bb9a/cert-manager-cainjector/0.log" Mar 21 00:31:20 crc kubenswrapper[5117]: I0321 00:31:20.165662 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-597b96b99b-7wclh_51772f05-28ae-4323-8633-c4996fb52222/cert-manager-webhook/0.log" Mar 21 00:31:28 crc kubenswrapper[5117]: I0321 00:31:28.238389 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-72hnj_c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e/kube-multus/0.log" Mar 21 00:31:28 crc kubenswrapper[5117]: I0321 00:31:28.238390 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-72hnj_c1ebb6fe-8a22-4bec-9cc2-97f1d02ec26e/kube-multus/0.log" Mar 21 00:31:28 crc kubenswrapper[5117]: I0321 00:31:28.256082 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:31:28 crc kubenswrapper[5117]: I0321 00:31:28.256126 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Mar 21 00:31:35 crc kubenswrapper[5117]: I0321 00:31:35.493806 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-55568fc96c-nv4sl_5f884179-adb4-4b62-bec6-14f40db13be5/prometheus-operator/0.log" Mar 21 00:31:35 crc kubenswrapper[5117]: I0321 00:31:35.528413 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5866674d46-mggcv_47f630cc-5d71-4431-8dea-397bfd008097/prometheus-operator-admission-webhook/0.log" Mar 21 00:31:35 crc kubenswrapper[5117]: I0321 00:31:35.673445 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5866674d46-pvrmj_68bdcac8-762b-4e94-8494-b571a8622dd8/prometheus-operator-admission-webhook/0.log" Mar 21 00:31:35 crc kubenswrapper[5117]: I0321 00:31:35.702520 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-587f9c8867-6xfv5_61a55f96-3dbb-43c7-9749-71d00e304e57/operator/0.log" Mar 21 00:31:35 crc kubenswrapper[5117]: I0321 00:31:35.841230 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-b577b4f9-4xhsq_4e836407-e9f1-4125-9810-741ae3b23431/perses-operator/0.log" Mar 21 00:31:51 crc kubenswrapper[5117]: I0321 00:31:51.422125 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v_364d7a25-35e5-49c8-b04a-79706601c454/util/0.log" Mar 21 00:31:51 crc kubenswrapper[5117]: I0321 00:31:51.562550 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v_364d7a25-35e5-49c8-b04a-79706601c454/util/0.log" Mar 21 00:31:51 crc kubenswrapper[5117]: I0321 00:31:51.566116 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v_364d7a25-35e5-49c8-b04a-79706601c454/pull/0.log" Mar 21 00:31:51 crc kubenswrapper[5117]: I0321 00:31:51.566126 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v_364d7a25-35e5-49c8-b04a-79706601c454/pull/0.log" Mar 21 00:31:51 crc kubenswrapper[5117]: I0321 00:31:51.724841 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v_364d7a25-35e5-49c8-b04a-79706601c454/util/0.log" Mar 21 00:31:51 crc kubenswrapper[5117]: I0321 00:31:51.778284 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v_364d7a25-35e5-49c8-b04a-79706601c454/extract/0.log" Mar 21 00:31:51 crc kubenswrapper[5117]: I0321 00:31:51.790680 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fg485v_364d7a25-35e5-49c8-b04a-79706601c454/pull/0.log" Mar 21 00:31:51 crc kubenswrapper[5117]: I0321 00:31:51.910403 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q_1cc8a25c-81b3-4ddf-bf36-ccbf077722be/util/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.080025 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q_1cc8a25c-81b3-4ddf-bf36-ccbf077722be/pull/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.108138 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q_1cc8a25c-81b3-4ddf-bf36-ccbf077722be/util/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.119236 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q_1cc8a25c-81b3-4ddf-bf36-ccbf077722be/pull/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.281104 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q_1cc8a25c-81b3-4ddf-bf36-ccbf077722be/util/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.303532 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q_1cc8a25c-81b3-4ddf-bf36-ccbf077722be/extract/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.303873 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39ek9s2q_1cc8a25c-81b3-4ddf-bf36-ccbf077722be/pull/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.461105 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st_79ef4008-b3a9-4f54-93b2-c296b982c5e7/util/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.627326 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st_79ef4008-b3a9-4f54-93b2-c296b982c5e7/util/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.658155 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st_79ef4008-b3a9-4f54-93b2-c296b982c5e7/pull/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.712132 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st_79ef4008-b3a9-4f54-93b2-c296b982c5e7/pull/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.799992 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st_79ef4008-b3a9-4f54-93b2-c296b982c5e7/util/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.837335 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st_79ef4008-b3a9-4f54-93b2-c296b982c5e7/extract/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.872592 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56l9st_79ef4008-b3a9-4f54-93b2-c296b982c5e7/pull/0.log" Mar 21 00:31:52 crc kubenswrapper[5117]: I0321 00:31:52.992925 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh_dc83e427-aef5-4984-b9e0-72d49d75c064/util/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.123762 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh_dc83e427-aef5-4984-b9e0-72d49d75c064/pull/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.155335 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh_dc83e427-aef5-4984-b9e0-72d49d75c064/util/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.221850 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh_dc83e427-aef5-4984-b9e0-72d49d75c064/pull/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.339398 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh_dc83e427-aef5-4984-b9e0-72d49d75c064/pull/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.341109 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh_dc83e427-aef5-4984-b9e0-72d49d75c064/extract/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.347164 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726nvgvh_dc83e427-aef5-4984-b9e0-72d49d75c064/util/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.555591 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v428c_879927af-6b88-4da1-915a-91728a3f2bd1/extract-utilities/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.694291 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v428c_879927af-6b88-4da1-915a-91728a3f2bd1/extract-utilities/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.710183 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v428c_879927af-6b88-4da1-915a-91728a3f2bd1/extract-content/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.716674 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v428c_879927af-6b88-4da1-915a-91728a3f2bd1/extract-content/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.956540 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v428c_879927af-6b88-4da1-915a-91728a3f2bd1/extract-utilities/0.log" Mar 21 00:31:53 crc kubenswrapper[5117]: I0321 00:31:53.963228 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v428c_879927af-6b88-4da1-915a-91728a3f2bd1/extract-content/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.133369 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hcbrw_ce1002d1-e98d-4caa-8a20-b1207c6f3ecd/extract-utilities/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.147112 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v428c_879927af-6b88-4da1-915a-91728a3f2bd1/registry-server/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.238374 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hcbrw_ce1002d1-e98d-4caa-8a20-b1207c6f3ecd/extract-content/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.271390 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hcbrw_ce1002d1-e98d-4caa-8a20-b1207c6f3ecd/extract-content/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.279296 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hcbrw_ce1002d1-e98d-4caa-8a20-b1207c6f3ecd/extract-utilities/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.406576 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hcbrw_ce1002d1-e98d-4caa-8a20-b1207c6f3ecd/extract-content/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.425455 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hcbrw_ce1002d1-e98d-4caa-8a20-b1207c6f3ecd/extract-utilities/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.512763 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-547dbd544d-p5w88_5beaa12a-d1ce-456a-a6f4-2d9b610c540f/marketplace-operator/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.603933 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-hcbrw_ce1002d1-e98d-4caa-8a20-b1207c6f3ecd/registry-server/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.677600 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6qtzc_38c8a935-8d54-4f0d-acb0-8ce00c78a95e/extract-utilities/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.766298 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6qtzc_38c8a935-8d54-4f0d-acb0-8ce00c78a95e/extract-content/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.772032 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6qtzc_38c8a935-8d54-4f0d-acb0-8ce00c78a95e/extract-utilities/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.796679 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6qtzc_38c8a935-8d54-4f0d-acb0-8ce00c78a95e/extract-content/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.925212 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6qtzc_38c8a935-8d54-4f0d-acb0-8ce00c78a95e/extract-content/0.log" Mar 21 00:31:54 crc kubenswrapper[5117]: I0321 00:31:54.938024 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6qtzc_38c8a935-8d54-4f0d-acb0-8ce00c78a95e/extract-utilities/0.log" Mar 21 00:31:55 crc kubenswrapper[5117]: I0321 00:31:55.089543 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6qtzc_38c8a935-8d54-4f0d-acb0-8ce00c78a95e/registry-server/0.log" Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.139890 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567552-gzz6x"] Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.150514 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567552-gzz6x" Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.151590 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567552-gzz6x"] Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.152952 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.155099 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.155380 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.343294 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7phc\" (UniqueName: \"kubernetes.io/projected/5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec-kube-api-access-j7phc\") pod \"auto-csr-approver-29567552-gzz6x\" (UID: \"5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec\") " pod="openshift-infra/auto-csr-approver-29567552-gzz6x" Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.444753 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-j7phc\" (UniqueName: \"kubernetes.io/projected/5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec-kube-api-access-j7phc\") pod \"auto-csr-approver-29567552-gzz6x\" (UID: \"5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec\") " pod="openshift-infra/auto-csr-approver-29567552-gzz6x" Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.477187 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7phc\" (UniqueName: \"kubernetes.io/projected/5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec-kube-api-access-j7phc\") pod \"auto-csr-approver-29567552-gzz6x\" (UID: \"5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec\") " pod="openshift-infra/auto-csr-approver-29567552-gzz6x" Mar 21 00:32:00 crc kubenswrapper[5117]: I0321 00:32:00.496909 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567552-gzz6x" Mar 21 00:32:01 crc kubenswrapper[5117]: I0321 00:32:01.035985 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567552-gzz6x"] Mar 21 00:32:01 crc kubenswrapper[5117]: I0321 00:32:01.547396 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567552-gzz6x" event={"ID":"5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec","Type":"ContainerStarted","Data":"83c528ec8bda1d974c6a3cb3c8df3ed97d884ecc4db8d6cf0609e5a0cccc5227"} Mar 21 00:32:02 crc kubenswrapper[5117]: I0321 00:32:02.561217 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567552-gzz6x" event={"ID":"5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec","Type":"ContainerStarted","Data":"a6d38552262fedd657a8c225109bbcbdc41497bcddb685edfba021193a944599"} Mar 21 00:32:02 crc kubenswrapper[5117]: I0321 00:32:02.581550 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29567552-gzz6x" podStartSLOduration=1.6977382049999998 podStartE2EDuration="2.581533701s" podCreationTimestamp="2026-03-21 00:32:00 +0000 UTC" firstStartedPulling="2026-03-21 00:32:01.04449321 +0000 UTC m=+1234.337780382" lastFinishedPulling="2026-03-21 00:32:01.928288666 +0000 UTC m=+1235.221575878" observedRunningTime="2026-03-21 00:32:02.580296852 +0000 UTC m=+1235.873584024" watchObservedRunningTime="2026-03-21 00:32:02.581533701 +0000 UTC m=+1235.874820873" Mar 21 00:32:03 crc kubenswrapper[5117]: I0321 00:32:03.571292 5117 generic.go:358] "Generic (PLEG): container finished" podID="5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec" containerID="a6d38552262fedd657a8c225109bbcbdc41497bcddb685edfba021193a944599" exitCode=0 Mar 21 00:32:03 crc kubenswrapper[5117]: I0321 00:32:03.571413 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567552-gzz6x" event={"ID":"5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec","Type":"ContainerDied","Data":"a6d38552262fedd657a8c225109bbcbdc41497bcddb685edfba021193a944599"} Mar 21 00:32:04 crc kubenswrapper[5117]: I0321 00:32:04.874700 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567552-gzz6x" Mar 21 00:32:04 crc kubenswrapper[5117]: I0321 00:32:04.953245 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7phc\" (UniqueName: \"kubernetes.io/projected/5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec-kube-api-access-j7phc\") pod \"5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec\" (UID: \"5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec\") " Mar 21 00:32:04 crc kubenswrapper[5117]: I0321 00:32:04.963950 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec-kube-api-access-j7phc" (OuterVolumeSpecName: "kube-api-access-j7phc") pod "5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec" (UID: "5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec"). InnerVolumeSpecName "kube-api-access-j7phc". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:32:05 crc kubenswrapper[5117]: I0321 00:32:05.055359 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-j7phc\" (UniqueName: \"kubernetes.io/projected/5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec-kube-api-access-j7phc\") on node \"crc\" DevicePath \"\"" Mar 21 00:32:05 crc kubenswrapper[5117]: I0321 00:32:05.588625 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567552-gzz6x" Mar 21 00:32:05 crc kubenswrapper[5117]: I0321 00:32:05.588717 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567552-gzz6x" event={"ID":"5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec","Type":"ContainerDied","Data":"83c528ec8bda1d974c6a3cb3c8df3ed97d884ecc4db8d6cf0609e5a0cccc5227"} Mar 21 00:32:05 crc kubenswrapper[5117]: I0321 00:32:05.589194 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83c528ec8bda1d974c6a3cb3c8df3ed97d884ecc4db8d6cf0609e5a0cccc5227" Mar 21 00:32:05 crc kubenswrapper[5117]: I0321 00:32:05.646729 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567546-f7pnm"] Mar 21 00:32:05 crc kubenswrapper[5117]: I0321 00:32:05.653243 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567546-f7pnm"] Mar 21 00:32:07 crc kubenswrapper[5117]: I0321 00:32:07.609984 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76525abe-9640-42c1-9871-bab8a511ddf0" path="/var/lib/kubelet/pods/76525abe-9640-42c1-9871-bab8a511ddf0/volumes" Mar 21 00:32:08 crc kubenswrapper[5117]: I0321 00:32:08.900959 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5866674d46-mggcv_47f630cc-5d71-4431-8dea-397bfd008097/prometheus-operator-admission-webhook/0.log" Mar 21 00:32:08 crc kubenswrapper[5117]: I0321 00:32:08.905536 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-55568fc96c-nv4sl_5f884179-adb4-4b62-bec6-14f40db13be5/prometheus-operator/0.log" Mar 21 00:32:08 crc kubenswrapper[5117]: I0321 00:32:08.941459 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5866674d46-pvrmj_68bdcac8-762b-4e94-8494-b571a8622dd8/prometheus-operator-admission-webhook/0.log" Mar 21 00:32:09 crc kubenswrapper[5117]: I0321 00:32:09.058835 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-b577b4f9-4xhsq_4e836407-e9f1-4125-9810-741ae3b23431/perses-operator/0.log" Mar 21 00:32:09 crc kubenswrapper[5117]: I0321 00:32:09.067526 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-587f9c8867-6xfv5_61a55f96-3dbb-43c7-9749-71d00e304e57/operator/0.log" Mar 21 00:32:31 crc kubenswrapper[5117]: I0321 00:32:31.178178 5117 scope.go:117] "RemoveContainer" containerID="2a14a9daae14497a58fa06ab75253f0f3105d56bd9250653366c77b511c09b3e" Mar 21 00:32:49 crc kubenswrapper[5117]: I0321 00:32:49.049645 5117 generic.go:358] "Generic (PLEG): container finished" podID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" containerID="5568cca41b73321c702ac54da1df361182483319e5c8df59523d75b1ce174c4f" exitCode=0 Mar 21 00:32:49 crc kubenswrapper[5117]: I0321 00:32:49.049797 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jj5pz/must-gather-vfn22" event={"ID":"6afcbed0-7a3d-4a2a-be56-de45f60b34d4","Type":"ContainerDied","Data":"5568cca41b73321c702ac54da1df361182483319e5c8df59523d75b1ce174c4f"} Mar 21 00:32:49 crc kubenswrapper[5117]: I0321 00:32:49.051381 5117 scope.go:117] "RemoveContainer" containerID="5568cca41b73321c702ac54da1df361182483319e5c8df59523d75b1ce174c4f" Mar 21 00:32:49 crc kubenswrapper[5117]: I0321 00:32:49.404473 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jj5pz_must-gather-vfn22_6afcbed0-7a3d-4a2a-be56-de45f60b34d4/gather/0.log" Mar 21 00:32:54 crc kubenswrapper[5117]: I0321 00:32:54.719062 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:32:54 crc kubenswrapper[5117]: I0321 00:32:54.719816 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:32:55 crc kubenswrapper[5117]: I0321 00:32:55.697617 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jj5pz/must-gather-vfn22"] Mar 21 00:32:55 crc kubenswrapper[5117]: I0321 00:32:55.698610 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-must-gather-jj5pz/must-gather-vfn22" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" containerName="copy" containerID="cri-o://5f46110e5afd0ae54d546a7780fc58d1f2fbca2a603e7d155817599c9af7c669" gracePeriod=2 Mar 21 00:32:55 crc kubenswrapper[5117]: I0321 00:32:55.701835 5117 status_manager.go:895] "Failed to get status for pod" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" pod="openshift-must-gather-jj5pz/must-gather-vfn22" err="pods \"must-gather-vfn22\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-jj5pz\": no relationship found between node 'crc' and this object" Mar 21 00:32:55 crc kubenswrapper[5117]: I0321 00:32:55.710730 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jj5pz/must-gather-vfn22"] Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.117473 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jj5pz_must-gather-vfn22_6afcbed0-7a3d-4a2a-be56-de45f60b34d4/copy/0.log" Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.118794 5117 generic.go:358] "Generic (PLEG): container finished" podID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" containerID="5f46110e5afd0ae54d546a7780fc58d1f2fbca2a603e7d155817599c9af7c669" exitCode=143 Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.118849 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db38b2702cdcd66638b544ddc9918b6f39c1eb4a2cea7f835306abc6f67867e7" Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.147771 5117 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jj5pz_must-gather-vfn22_6afcbed0-7a3d-4a2a-be56-de45f60b34d4/copy/0.log" Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.148436 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.149696 5117 status_manager.go:895] "Failed to get status for pod" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" pod="openshift-must-gather-jj5pz/must-gather-vfn22" err="pods \"must-gather-vfn22\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-jj5pz\": no relationship found between node 'crc' and this object" Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.199295 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vngcs\" (UniqueName: \"kubernetes.io/projected/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-kube-api-access-vngcs\") pod \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\" (UID: \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\") " Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.199409 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-must-gather-output\") pod \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\" (UID: \"6afcbed0-7a3d-4a2a-be56-de45f60b34d4\") " Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.209462 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-kube-api-access-vngcs" (OuterVolumeSpecName: "kube-api-access-vngcs") pod "6afcbed0-7a3d-4a2a-be56-de45f60b34d4" (UID: "6afcbed0-7a3d-4a2a-be56-de45f60b34d4"). InnerVolumeSpecName "kube-api-access-vngcs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.274015 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6afcbed0-7a3d-4a2a-be56-de45f60b34d4" (UID: "6afcbed0-7a3d-4a2a-be56-de45f60b34d4"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.303000 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vngcs\" (UniqueName: \"kubernetes.io/projected/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-kube-api-access-vngcs\") on node \"crc\" DevicePath \"\"" Mar 21 00:32:56 crc kubenswrapper[5117]: I0321 00:32:56.303041 5117 reconciler_common.go:299] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6afcbed0-7a3d-4a2a-be56-de45f60b34d4-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 21 00:32:57 crc kubenswrapper[5117]: I0321 00:32:57.127291 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jj5pz/must-gather-vfn22" Mar 21 00:32:57 crc kubenswrapper[5117]: I0321 00:32:57.129622 5117 status_manager.go:895] "Failed to get status for pod" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" pod="openshift-must-gather-jj5pz/must-gather-vfn22" err="pods \"must-gather-vfn22\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-jj5pz\": no relationship found between node 'crc' and this object" Mar 21 00:32:57 crc kubenswrapper[5117]: I0321 00:32:57.155848 5117 status_manager.go:895] "Failed to get status for pod" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" pod="openshift-must-gather-jj5pz/must-gather-vfn22" err="pods \"must-gather-vfn22\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-jj5pz\": no relationship found between node 'crc' and this object" Mar 21 00:32:57 crc kubenswrapper[5117]: I0321 00:32:57.591502 5117 status_manager.go:895] "Failed to get status for pod" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" pod="openshift-must-gather-jj5pz/must-gather-vfn22" err="pods \"must-gather-vfn22\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-jj5pz\": no relationship found between node 'crc' and this object" Mar 21 00:32:57 crc kubenswrapper[5117]: I0321 00:32:57.591765 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" path="/var/lib/kubelet/pods/6afcbed0-7a3d-4a2a-be56-de45f60b34d4/volumes" Mar 21 00:33:24 crc kubenswrapper[5117]: I0321 00:33:24.718999 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:33:24 crc kubenswrapper[5117]: I0321 00:33:24.719677 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:33:54 crc kubenswrapper[5117]: I0321 00:33:54.719019 5117 patch_prober.go:28] interesting pod/machine-config-daemon-w9wqq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 21 00:33:54 crc kubenswrapper[5117]: I0321 00:33:54.719742 5117 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 21 00:33:54 crc kubenswrapper[5117]: I0321 00:33:54.719811 5117 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" Mar 21 00:33:54 crc kubenswrapper[5117]: I0321 00:33:54.720624 5117 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7bf5d1dd104623548e3c0f4587a917340dbfe836ba859055e5a1a9dbd9072797"} pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 21 00:33:54 crc kubenswrapper[5117]: I0321 00:33:54.720758 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" podUID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerName="machine-config-daemon" containerID="cri-o://7bf5d1dd104623548e3c0f4587a917340dbfe836ba859055e5a1a9dbd9072797" gracePeriod=600 Mar 21 00:33:54 crc kubenswrapper[5117]: I0321 00:33:54.861799 5117 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 21 00:33:55 crc kubenswrapper[5117]: I0321 00:33:55.688997 5117 generic.go:358] "Generic (PLEG): container finished" podID="6cb44ea3-ded3-4e25-a367-c6b9f0ab2153" containerID="7bf5d1dd104623548e3c0f4587a917340dbfe836ba859055e5a1a9dbd9072797" exitCode=0 Mar 21 00:33:55 crc kubenswrapper[5117]: I0321 00:33:55.689056 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerDied","Data":"7bf5d1dd104623548e3c0f4587a917340dbfe836ba859055e5a1a9dbd9072797"} Mar 21 00:33:55 crc kubenswrapper[5117]: I0321 00:33:55.689440 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9wqq" event={"ID":"6cb44ea3-ded3-4e25-a367-c6b9f0ab2153","Type":"ContainerStarted","Data":"5b69dd4eae7deb67718bd4a3286476749a1718205d300d9a17c53ef57f42d4a5"} Mar 21 00:33:55 crc kubenswrapper[5117]: I0321 00:33:55.689476 5117 scope.go:117] "RemoveContainer" containerID="eb586c1751d61bfe9aa828e1aeabf26cbd66967a6d89d22ac2c57498df26a314" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.148101 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29567554-jjk7j"] Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.150185 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" containerName="gather" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.150317 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" containerName="gather" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.150428 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" containerName="copy" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.150515 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" containerName="copy" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.150587 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec" containerName="oc" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.150662 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec" containerName="oc" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.150926 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" containerName="copy" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.151015 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="5ba9130a-2192-41f4-a1e4-5a3e61a1f1ec" containerName="oc" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.151091 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="6afcbed0-7a3d-4a2a-be56-de45f60b34d4" containerName="gather" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.156522 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567554-jjk7j" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.158423 5117 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-infra\"/\"csr-approver-sa-dockercfg-9hl8s\"" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.160201 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"openshift-service-ca.crt\"" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.160309 5117 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-infra\"/\"kube-root-ca.crt\"" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.169477 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567554-jjk7j"] Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.325197 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v2tn\" (UniqueName: \"kubernetes.io/projected/1691b555-c354-4b2b-9557-7af912d1b3b7-kube-api-access-5v2tn\") pod \"auto-csr-approver-29567554-jjk7j\" (UID: \"1691b555-c354-4b2b-9557-7af912d1b3b7\") " pod="openshift-infra/auto-csr-approver-29567554-jjk7j" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.426913 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5v2tn\" (UniqueName: \"kubernetes.io/projected/1691b555-c354-4b2b-9557-7af912d1b3b7-kube-api-access-5v2tn\") pod \"auto-csr-approver-29567554-jjk7j\" (UID: \"1691b555-c354-4b2b-9557-7af912d1b3b7\") " pod="openshift-infra/auto-csr-approver-29567554-jjk7j" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.461123 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v2tn\" (UniqueName: \"kubernetes.io/projected/1691b555-c354-4b2b-9557-7af912d1b3b7-kube-api-access-5v2tn\") pod \"auto-csr-approver-29567554-jjk7j\" (UID: \"1691b555-c354-4b2b-9557-7af912d1b3b7\") " pod="openshift-infra/auto-csr-approver-29567554-jjk7j" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.485292 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567554-jjk7j" Mar 21 00:34:00 crc kubenswrapper[5117]: I0321 00:34:00.743018 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29567554-jjk7j"] Mar 21 00:34:01 crc kubenswrapper[5117]: I0321 00:34:01.754770 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567554-jjk7j" event={"ID":"1691b555-c354-4b2b-9557-7af912d1b3b7","Type":"ContainerStarted","Data":"3543dde8043535b54218ac5cde96aff556301a8a382573d5387756bbea3499e6"} Mar 21 00:34:02 crc kubenswrapper[5117]: I0321 00:34:02.767343 5117 generic.go:358] "Generic (PLEG): container finished" podID="1691b555-c354-4b2b-9557-7af912d1b3b7" containerID="9af68d556b61de9d0aaff6af997b73f0e3ce743d12e45f2efe9acb9b59228292" exitCode=0 Mar 21 00:34:02 crc kubenswrapper[5117]: I0321 00:34:02.767438 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567554-jjk7j" event={"ID":"1691b555-c354-4b2b-9557-7af912d1b3b7","Type":"ContainerDied","Data":"9af68d556b61de9d0aaff6af997b73f0e3ce743d12e45f2efe9acb9b59228292"} Mar 21 00:34:04 crc kubenswrapper[5117]: I0321 00:34:04.097397 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567554-jjk7j" Mar 21 00:34:04 crc kubenswrapper[5117]: I0321 00:34:04.288173 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v2tn\" (UniqueName: \"kubernetes.io/projected/1691b555-c354-4b2b-9557-7af912d1b3b7-kube-api-access-5v2tn\") pod \"1691b555-c354-4b2b-9557-7af912d1b3b7\" (UID: \"1691b555-c354-4b2b-9557-7af912d1b3b7\") " Mar 21 00:34:04 crc kubenswrapper[5117]: I0321 00:34:04.301656 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1691b555-c354-4b2b-9557-7af912d1b3b7-kube-api-access-5v2tn" (OuterVolumeSpecName: "kube-api-access-5v2tn") pod "1691b555-c354-4b2b-9557-7af912d1b3b7" (UID: "1691b555-c354-4b2b-9557-7af912d1b3b7"). InnerVolumeSpecName "kube-api-access-5v2tn". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:34:04 crc kubenswrapper[5117]: I0321 00:34:04.390367 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-5v2tn\" (UniqueName: \"kubernetes.io/projected/1691b555-c354-4b2b-9557-7af912d1b3b7-kube-api-access-5v2tn\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:04 crc kubenswrapper[5117]: I0321 00:34:04.790516 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29567554-jjk7j" event={"ID":"1691b555-c354-4b2b-9557-7af912d1b3b7","Type":"ContainerDied","Data":"3543dde8043535b54218ac5cde96aff556301a8a382573d5387756bbea3499e6"} Mar 21 00:34:04 crc kubenswrapper[5117]: I0321 00:34:04.791477 5117 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3543dde8043535b54218ac5cde96aff556301a8a382573d5387756bbea3499e6" Mar 21 00:34:04 crc kubenswrapper[5117]: I0321 00:34:04.790602 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29567554-jjk7j" Mar 21 00:34:05 crc kubenswrapper[5117]: I0321 00:34:05.173474 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29567548-gfvwn"] Mar 21 00:34:05 crc kubenswrapper[5117]: I0321 00:34:05.185305 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29567548-gfvwn"] Mar 21 00:34:05 crc kubenswrapper[5117]: I0321 00:34:05.591642 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b31000c-6a98-4e25-95cd-510fabb91ae2" path="/var/lib/kubelet/pods/8b31000c-6a98-4e25-95cd-510fabb91ae2/volumes" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.665281 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jmtwg"] Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.666849 5117 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1691b555-c354-4b2b-9557-7af912d1b3b7" containerName="oc" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.666866 5117 state_mem.go:107] "Deleted CPUSet assignment" podUID="1691b555-c354-4b2b-9557-7af912d1b3b7" containerName="oc" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.667048 5117 memory_manager.go:356] "RemoveStaleState removing state" podUID="1691b555-c354-4b2b-9557-7af912d1b3b7" containerName="oc" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.681300 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jmtwg"] Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.681478 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.785305 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-catalog-content\") pod \"certified-operators-jmtwg\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.785929 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z868r\" (UniqueName: \"kubernetes.io/projected/a57ed631-177f-4f91-9667-7ad813439767-kube-api-access-z868r\") pod \"certified-operators-jmtwg\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.786047 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-utilities\") pod \"certified-operators-jmtwg\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.857134 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4wng8"] Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.862152 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.884394 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wng8"] Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.887872 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-z868r\" (UniqueName: \"kubernetes.io/projected/a57ed631-177f-4f91-9667-7ad813439767-kube-api-access-z868r\") pod \"certified-operators-jmtwg\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.887969 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-utilities\") pod \"certified-operators-jmtwg\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.888319 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-catalog-content\") pod \"certified-operators-jmtwg\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.888938 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-utilities\") pod \"certified-operators-jmtwg\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.889085 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-catalog-content\") pod \"certified-operators-jmtwg\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.919776 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-z868r\" (UniqueName: \"kubernetes.io/projected/a57ed631-177f-4f91-9667-7ad813439767-kube-api-access-z868r\") pod \"certified-operators-jmtwg\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.989591 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-utilities\") pod \"redhat-operators-4wng8\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.989959 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-catalog-content\") pod \"redhat-operators-4wng8\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:15 crc kubenswrapper[5117]: I0321 00:34:15.989999 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vjkr\" (UniqueName: \"kubernetes.io/projected/90eec1f2-94c4-4330-be73-bbb38d7bfad7-kube-api-access-7vjkr\") pod \"redhat-operators-4wng8\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.008525 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.091471 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-utilities\") pod \"redhat-operators-4wng8\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.091682 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-catalog-content\") pod \"redhat-operators-4wng8\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.091795 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7vjkr\" (UniqueName: \"kubernetes.io/projected/90eec1f2-94c4-4330-be73-bbb38d7bfad7-kube-api-access-7vjkr\") pod \"redhat-operators-4wng8\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.091986 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-utilities\") pod \"redhat-operators-4wng8\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.092391 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-catalog-content\") pod \"redhat-operators-4wng8\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.143989 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vjkr\" (UniqueName: \"kubernetes.io/projected/90eec1f2-94c4-4330-be73-bbb38d7bfad7-kube-api-access-7vjkr\") pod \"redhat-operators-4wng8\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.189284 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.263810 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jmtwg"] Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.623606 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wng8"] Mar 21 00:34:16 crc kubenswrapper[5117]: W0321 00:34:16.627463 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90eec1f2_94c4_4330_be73_bbb38d7bfad7.slice/crio-2f06d1f3b197596570a0d07d6d1c4c29d78e749c421ef21a775c5c5a0ef33561 WatchSource:0}: Error finding container 2f06d1f3b197596570a0d07d6d1c4c29d78e749c421ef21a775c5c5a0ef33561: Status 404 returned error can't find the container with id 2f06d1f3b197596570a0d07d6d1c4c29d78e749c421ef21a775c5c5a0ef33561 Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.905021 5117 generic.go:358] "Generic (PLEG): container finished" podID="a57ed631-177f-4f91-9667-7ad813439767" containerID="8fee967268041d4d231ae44babeb7e81b10b22225c93f316bfde8353a1dfdf38" exitCode=0 Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.905117 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmtwg" event={"ID":"a57ed631-177f-4f91-9667-7ad813439767","Type":"ContainerDied","Data":"8fee967268041d4d231ae44babeb7e81b10b22225c93f316bfde8353a1dfdf38"} Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.905451 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmtwg" event={"ID":"a57ed631-177f-4f91-9667-7ad813439767","Type":"ContainerStarted","Data":"8cd8932108d06983f4ce6b0618a836aa43a8050a545adf762b51fcb3e2cbf10e"} Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.909070 5117 generic.go:358] "Generic (PLEG): container finished" podID="90eec1f2-94c4-4330-be73-bbb38d7bfad7" containerID="202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f" exitCode=0 Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.909104 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wng8" event={"ID":"90eec1f2-94c4-4330-be73-bbb38d7bfad7","Type":"ContainerDied","Data":"202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f"} Mar 21 00:34:16 crc kubenswrapper[5117]: I0321 00:34:16.909167 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wng8" event={"ID":"90eec1f2-94c4-4330-be73-bbb38d7bfad7","Type":"ContainerStarted","Data":"2f06d1f3b197596570a0d07d6d1c4c29d78e749c421ef21a775c5c5a0ef33561"} Mar 21 00:34:17 crc kubenswrapper[5117]: I0321 00:34:17.919963 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmtwg" event={"ID":"a57ed631-177f-4f91-9667-7ad813439767","Type":"ContainerStarted","Data":"24cc18ab29265a8a74a0f3d784ed38830828079089d0d65adce164a91b634578"} Mar 21 00:34:17 crc kubenswrapper[5117]: I0321 00:34:17.923673 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wng8" event={"ID":"90eec1f2-94c4-4330-be73-bbb38d7bfad7","Type":"ContainerStarted","Data":"4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1"} Mar 21 00:34:18 crc kubenswrapper[5117]: I0321 00:34:18.933345 5117 generic.go:358] "Generic (PLEG): container finished" podID="90eec1f2-94c4-4330-be73-bbb38d7bfad7" containerID="4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1" exitCode=0 Mar 21 00:34:18 crc kubenswrapper[5117]: I0321 00:34:18.933454 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wng8" event={"ID":"90eec1f2-94c4-4330-be73-bbb38d7bfad7","Type":"ContainerDied","Data":"4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1"} Mar 21 00:34:18 crc kubenswrapper[5117]: I0321 00:34:18.936131 5117 generic.go:358] "Generic (PLEG): container finished" podID="a57ed631-177f-4f91-9667-7ad813439767" containerID="24cc18ab29265a8a74a0f3d784ed38830828079089d0d65adce164a91b634578" exitCode=0 Mar 21 00:34:18 crc kubenswrapper[5117]: I0321 00:34:18.936196 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmtwg" event={"ID":"a57ed631-177f-4f91-9667-7ad813439767","Type":"ContainerDied","Data":"24cc18ab29265a8a74a0f3d784ed38830828079089d0d65adce164a91b634578"} Mar 21 00:34:19 crc kubenswrapper[5117]: I0321 00:34:19.949274 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmtwg" event={"ID":"a57ed631-177f-4f91-9667-7ad813439767","Type":"ContainerStarted","Data":"edf0160a9f69c9a6e79b741c1305b7aefa59dbe67471d155d6de25f76f31895c"} Mar 21 00:34:19 crc kubenswrapper[5117]: I0321 00:34:19.953557 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wng8" event={"ID":"90eec1f2-94c4-4330-be73-bbb38d7bfad7","Type":"ContainerStarted","Data":"45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578"} Mar 21 00:34:19 crc kubenswrapper[5117]: I0321 00:34:19.978148 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jmtwg" podStartSLOduration=4.267446911 podStartE2EDuration="4.978128545s" podCreationTimestamp="2026-03-21 00:34:15 +0000 UTC" firstStartedPulling="2026-03-21 00:34:16.907020976 +0000 UTC m=+1370.200308148" lastFinishedPulling="2026-03-21 00:34:17.61770261 +0000 UTC m=+1370.910989782" observedRunningTime="2026-03-21 00:34:19.975701061 +0000 UTC m=+1373.268988303" watchObservedRunningTime="2026-03-21 00:34:19.978128545 +0000 UTC m=+1373.271415717" Mar 21 00:34:19 crc kubenswrapper[5117]: I0321 00:34:19.999541 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4wng8" podStartSLOduration=4.363162869 podStartE2EDuration="4.99952067s" podCreationTimestamp="2026-03-21 00:34:15 +0000 UTC" firstStartedPulling="2026-03-21 00:34:16.909830282 +0000 UTC m=+1370.203117454" lastFinishedPulling="2026-03-21 00:34:17.546188043 +0000 UTC m=+1370.839475255" observedRunningTime="2026-03-21 00:34:19.993008881 +0000 UTC m=+1373.286296063" watchObservedRunningTime="2026-03-21 00:34:19.99952067 +0000 UTC m=+1373.292807842" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.457886 5117 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8xfpq"] Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.472212 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.480563 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8xfpq"] Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.564000 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-utilities\") pod \"community-operators-8xfpq\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.564058 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw5c8\" (UniqueName: \"kubernetes.io/projected/9bf33857-8eef-4622-8344-21e7a8d0382e-kube-api-access-pw5c8\") pod \"community-operators-8xfpq\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.564088 5117 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-catalog-content\") pod \"community-operators-8xfpq\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.665941 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pw5c8\" (UniqueName: \"kubernetes.io/projected/9bf33857-8eef-4622-8344-21e7a8d0382e-kube-api-access-pw5c8\") pod \"community-operators-8xfpq\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.665983 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-catalog-content\") pod \"community-operators-8xfpq\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.666099 5117 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-utilities\") pod \"community-operators-8xfpq\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.666526 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-utilities\") pod \"community-operators-8xfpq\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.667022 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-catalog-content\") pod \"community-operators-8xfpq\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.689792 5117 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw5c8\" (UniqueName: \"kubernetes.io/projected/9bf33857-8eef-4622-8344-21e7a8d0382e-kube-api-access-pw5c8\") pod \"community-operators-8xfpq\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:20 crc kubenswrapper[5117]: I0321 00:34:20.797718 5117 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:21 crc kubenswrapper[5117]: I0321 00:34:21.295775 5117 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8xfpq"] Mar 21 00:34:21 crc kubenswrapper[5117]: W0321 00:34:21.306595 5117 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bf33857_8eef_4622_8344_21e7a8d0382e.slice/crio-f8aa825dcfbb12ee8412b5344ae8964e39d7bf238a74b4042a97aadf192350c4 WatchSource:0}: Error finding container f8aa825dcfbb12ee8412b5344ae8964e39d7bf238a74b4042a97aadf192350c4: Status 404 returned error can't find the container with id f8aa825dcfbb12ee8412b5344ae8964e39d7bf238a74b4042a97aadf192350c4 Mar 21 00:34:21 crc kubenswrapper[5117]: I0321 00:34:21.975410 5117 generic.go:358] "Generic (PLEG): container finished" podID="9bf33857-8eef-4622-8344-21e7a8d0382e" containerID="62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb" exitCode=0 Mar 21 00:34:21 crc kubenswrapper[5117]: I0321 00:34:21.975505 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xfpq" event={"ID":"9bf33857-8eef-4622-8344-21e7a8d0382e","Type":"ContainerDied","Data":"62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb"} Mar 21 00:34:21 crc kubenswrapper[5117]: I0321 00:34:21.975583 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xfpq" event={"ID":"9bf33857-8eef-4622-8344-21e7a8d0382e","Type":"ContainerStarted","Data":"f8aa825dcfbb12ee8412b5344ae8964e39d7bf238a74b4042a97aadf192350c4"} Mar 21 00:34:22 crc kubenswrapper[5117]: I0321 00:34:22.984963 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xfpq" event={"ID":"9bf33857-8eef-4622-8344-21e7a8d0382e","Type":"ContainerStarted","Data":"3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516"} Mar 21 00:34:25 crc kubenswrapper[5117]: I0321 00:34:25.001839 5117 generic.go:358] "Generic (PLEG): container finished" podID="9bf33857-8eef-4622-8344-21e7a8d0382e" containerID="3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516" exitCode=0 Mar 21 00:34:25 crc kubenswrapper[5117]: I0321 00:34:25.001882 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xfpq" event={"ID":"9bf33857-8eef-4622-8344-21e7a8d0382e","Type":"ContainerDied","Data":"3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516"} Mar 21 00:34:26 crc kubenswrapper[5117]: I0321 00:34:26.008908 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:26 crc kubenswrapper[5117]: I0321 00:34:26.008954 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:26 crc kubenswrapper[5117]: I0321 00:34:26.013840 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xfpq" event={"ID":"9bf33857-8eef-4622-8344-21e7a8d0382e","Type":"ContainerStarted","Data":"e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974"} Mar 21 00:34:26 crc kubenswrapper[5117]: I0321 00:34:26.046838 5117 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8xfpq" podStartSLOduration=5.452986123 podStartE2EDuration="6.046807503s" podCreationTimestamp="2026-03-21 00:34:20 +0000 UTC" firstStartedPulling="2026-03-21 00:34:21.976981063 +0000 UTC m=+1375.270268265" lastFinishedPulling="2026-03-21 00:34:22.570802443 +0000 UTC m=+1375.864089645" observedRunningTime="2026-03-21 00:34:26.042290216 +0000 UTC m=+1379.335577398" watchObservedRunningTime="2026-03-21 00:34:26.046807503 +0000 UTC m=+1379.340094715" Mar 21 00:34:26 crc kubenswrapper[5117]: I0321 00:34:26.079366 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:26 crc kubenswrapper[5117]: I0321 00:34:26.189870 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:26 crc kubenswrapper[5117]: I0321 00:34:26.189934 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:26 crc kubenswrapper[5117]: I0321 00:34:26.257049 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:27 crc kubenswrapper[5117]: I0321 00:34:27.078300 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:27 crc kubenswrapper[5117]: I0321 00:34:27.078713 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:29 crc kubenswrapper[5117]: I0321 00:34:29.244381 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wng8"] Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.054340 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4wng8" podUID="90eec1f2-94c4-4330-be73-bbb38d7bfad7" containerName="registry-server" containerID="cri-o://45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578" gracePeriod=2 Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.522919 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.639486 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vjkr\" (UniqueName: \"kubernetes.io/projected/90eec1f2-94c4-4330-be73-bbb38d7bfad7-kube-api-access-7vjkr\") pod \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.639880 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-catalog-content\") pod \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.640017 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-utilities\") pod \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\" (UID: \"90eec1f2-94c4-4330-be73-bbb38d7bfad7\") " Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.642420 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-utilities" (OuterVolumeSpecName: "utilities") pod "90eec1f2-94c4-4330-be73-bbb38d7bfad7" (UID: "90eec1f2-94c4-4330-be73-bbb38d7bfad7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.649000 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90eec1f2-94c4-4330-be73-bbb38d7bfad7-kube-api-access-7vjkr" (OuterVolumeSpecName: "kube-api-access-7vjkr") pod "90eec1f2-94c4-4330-be73-bbb38d7bfad7" (UID: "90eec1f2-94c4-4330-be73-bbb38d7bfad7"). InnerVolumeSpecName "kube-api-access-7vjkr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.741567 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.741639 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7vjkr\" (UniqueName: \"kubernetes.io/projected/90eec1f2-94c4-4330-be73-bbb38d7bfad7-kube-api-access-7vjkr\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.763755 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90eec1f2-94c4-4330-be73-bbb38d7bfad7" (UID: "90eec1f2-94c4-4330-be73-bbb38d7bfad7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.798907 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.799019 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.843173 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90eec1f2-94c4-4330-be73-bbb38d7bfad7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:30 crc kubenswrapper[5117]: I0321 00:34:30.873062 5117 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.069782 5117 generic.go:358] "Generic (PLEG): container finished" podID="90eec1f2-94c4-4330-be73-bbb38d7bfad7" containerID="45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578" exitCode=0 Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.069907 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wng8" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.069898 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wng8" event={"ID":"90eec1f2-94c4-4330-be73-bbb38d7bfad7","Type":"ContainerDied","Data":"45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578"} Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.070083 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wng8" event={"ID":"90eec1f2-94c4-4330-be73-bbb38d7bfad7","Type":"ContainerDied","Data":"2f06d1f3b197596570a0d07d6d1c4c29d78e749c421ef21a775c5c5a0ef33561"} Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.070117 5117 scope.go:117] "RemoveContainer" containerID="45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.122413 5117 scope.go:117] "RemoveContainer" containerID="4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.127742 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wng8"] Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.137225 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4wng8"] Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.137615 5117 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.154624 5117 scope.go:117] "RemoveContainer" containerID="202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.179755 5117 scope.go:117] "RemoveContainer" containerID="45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578" Mar 21 00:34:31 crc kubenswrapper[5117]: E0321 00:34:31.180303 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578\": container with ID starting with 45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578 not found: ID does not exist" containerID="45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.180354 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578"} err="failed to get container status \"45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578\": rpc error: code = NotFound desc = could not find container \"45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578\": container with ID starting with 45019a7206710e4fa3de986cf5e4f5e372e28e8dfba780b78238bf72b3558578 not found: ID does not exist" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.180381 5117 scope.go:117] "RemoveContainer" containerID="4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1" Mar 21 00:34:31 crc kubenswrapper[5117]: E0321 00:34:31.180644 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1\": container with ID starting with 4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1 not found: ID does not exist" containerID="4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.180681 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1"} err="failed to get container status \"4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1\": rpc error: code = NotFound desc = could not find container \"4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1\": container with ID starting with 4e41ec9c05617d64f73d77a3d82b5327e6c8f3df076d1b332f85f88cf6c381d1 not found: ID does not exist" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.180704 5117 scope.go:117] "RemoveContainer" containerID="202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f" Mar 21 00:34:31 crc kubenswrapper[5117]: E0321 00:34:31.181169 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f\": container with ID starting with 202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f not found: ID does not exist" containerID="202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.181204 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f"} err="failed to get container status \"202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f\": rpc error: code = NotFound desc = could not find container \"202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f\": container with ID starting with 202dc4a2ac54592e4d469a0fa85a13dc7596a99c3119da91a1f2b7426c8c1e2f not found: ID does not exist" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.364240 5117 scope.go:117] "RemoveContainer" containerID="15c5368178c60d1eff91c02a3a2cf4bee8a77669b672a0d1b26ac19f961204e8" Mar 21 00:34:31 crc kubenswrapper[5117]: I0321 00:34:31.592214 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90eec1f2-94c4-4330-be73-bbb38d7bfad7" path="/var/lib/kubelet/pods/90eec1f2-94c4-4330-be73-bbb38d7bfad7/volumes" Mar 21 00:34:32 crc kubenswrapper[5117]: I0321 00:34:32.848929 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jmtwg"] Mar 21 00:34:32 crc kubenswrapper[5117]: I0321 00:34:32.850022 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jmtwg" podUID="a57ed631-177f-4f91-9667-7ad813439767" containerName="registry-server" containerID="cri-o://edf0160a9f69c9a6e79b741c1305b7aefa59dbe67471d155d6de25f76f31895c" gracePeriod=2 Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.095037 5117 generic.go:358] "Generic (PLEG): container finished" podID="a57ed631-177f-4f91-9667-7ad813439767" containerID="edf0160a9f69c9a6e79b741c1305b7aefa59dbe67471d155d6de25f76f31895c" exitCode=0 Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.095128 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmtwg" event={"ID":"a57ed631-177f-4f91-9667-7ad813439767","Type":"ContainerDied","Data":"edf0160a9f69c9a6e79b741c1305b7aefa59dbe67471d155d6de25f76f31895c"} Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.389576 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.492053 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-utilities\") pod \"a57ed631-177f-4f91-9667-7ad813439767\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.492191 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z868r\" (UniqueName: \"kubernetes.io/projected/a57ed631-177f-4f91-9667-7ad813439767-kube-api-access-z868r\") pod \"a57ed631-177f-4f91-9667-7ad813439767\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.492241 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-catalog-content\") pod \"a57ed631-177f-4f91-9667-7ad813439767\" (UID: \"a57ed631-177f-4f91-9667-7ad813439767\") " Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.494093 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-utilities" (OuterVolumeSpecName: "utilities") pod "a57ed631-177f-4f91-9667-7ad813439767" (UID: "a57ed631-177f-4f91-9667-7ad813439767"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.502947 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57ed631-177f-4f91-9667-7ad813439767-kube-api-access-z868r" (OuterVolumeSpecName: "kube-api-access-z868r") pod "a57ed631-177f-4f91-9667-7ad813439767" (UID: "a57ed631-177f-4f91-9667-7ad813439767"). InnerVolumeSpecName "kube-api-access-z868r". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.577977 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a57ed631-177f-4f91-9667-7ad813439767" (UID: "a57ed631-177f-4f91-9667-7ad813439767"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.596040 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.596095 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-z868r\" (UniqueName: \"kubernetes.io/projected/a57ed631-177f-4f91-9667-7ad813439767-kube-api-access-z868r\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:33 crc kubenswrapper[5117]: I0321 00:34:33.596116 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57ed631-177f-4f91-9667-7ad813439767-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.055667 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8xfpq"] Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.057174 5117 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8xfpq" podUID="9bf33857-8eef-4622-8344-21e7a8d0382e" containerName="registry-server" containerID="cri-o://e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974" gracePeriod=2 Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.112355 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jmtwg" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.112353 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jmtwg" event={"ID":"a57ed631-177f-4f91-9667-7ad813439767","Type":"ContainerDied","Data":"8cd8932108d06983f4ce6b0618a836aa43a8050a545adf762b51fcb3e2cbf10e"} Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.112584 5117 scope.go:117] "RemoveContainer" containerID="edf0160a9f69c9a6e79b741c1305b7aefa59dbe67471d155d6de25f76f31895c" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.268131 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jmtwg"] Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.275872 5117 scope.go:117] "RemoveContainer" containerID="24cc18ab29265a8a74a0f3d784ed38830828079089d0d65adce164a91b634578" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.276566 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jmtwg"] Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.333873 5117 scope.go:117] "RemoveContainer" containerID="8fee967268041d4d231ae44babeb7e81b10b22225c93f316bfde8353a1dfdf38" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.542224 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.613203 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-catalog-content\") pod \"9bf33857-8eef-4622-8344-21e7a8d0382e\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.613395 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-utilities\") pod \"9bf33857-8eef-4622-8344-21e7a8d0382e\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.613869 5117 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw5c8\" (UniqueName: \"kubernetes.io/projected/9bf33857-8eef-4622-8344-21e7a8d0382e-kube-api-access-pw5c8\") pod \"9bf33857-8eef-4622-8344-21e7a8d0382e\" (UID: \"9bf33857-8eef-4622-8344-21e7a8d0382e\") " Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.616540 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-utilities" (OuterVolumeSpecName: "utilities") pod "9bf33857-8eef-4622-8344-21e7a8d0382e" (UID: "9bf33857-8eef-4622-8344-21e7a8d0382e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.619690 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bf33857-8eef-4622-8344-21e7a8d0382e-kube-api-access-pw5c8" (OuterVolumeSpecName: "kube-api-access-pw5c8") pod "9bf33857-8eef-4622-8344-21e7a8d0382e" (UID: "9bf33857-8eef-4622-8344-21e7a8d0382e"). InnerVolumeSpecName "kube-api-access-pw5c8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.688184 5117 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bf33857-8eef-4622-8344-21e7a8d0382e" (UID: "9bf33857-8eef-4622-8344-21e7a8d0382e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.716566 5117 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pw5c8\" (UniqueName: \"kubernetes.io/projected/9bf33857-8eef-4622-8344-21e7a8d0382e-kube-api-access-pw5c8\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.716614 5117 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:34 crc kubenswrapper[5117]: I0321 00:34:34.716625 5117 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf33857-8eef-4622-8344-21e7a8d0382e-utilities\") on node \"crc\" DevicePath \"\"" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.126494 5117 generic.go:358] "Generic (PLEG): container finished" podID="9bf33857-8eef-4622-8344-21e7a8d0382e" containerID="e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974" exitCode=0 Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.126607 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xfpq" event={"ID":"9bf33857-8eef-4622-8344-21e7a8d0382e","Type":"ContainerDied","Data":"e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974"} Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.126630 5117 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8xfpq" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.126660 5117 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8xfpq" event={"ID":"9bf33857-8eef-4622-8344-21e7a8d0382e","Type":"ContainerDied","Data":"f8aa825dcfbb12ee8412b5344ae8964e39d7bf238a74b4042a97aadf192350c4"} Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.126685 5117 scope.go:117] "RemoveContainer" containerID="e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.163213 5117 scope.go:117] "RemoveContainer" containerID="3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.189530 5117 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8xfpq"] Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.197890 5117 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8xfpq"] Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.224975 5117 scope.go:117] "RemoveContainer" containerID="62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.261515 5117 scope.go:117] "RemoveContainer" containerID="e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974" Mar 21 00:34:35 crc kubenswrapper[5117]: E0321 00:34:35.262068 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974\": container with ID starting with e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974 not found: ID does not exist" containerID="e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.262122 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974"} err="failed to get container status \"e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974\": rpc error: code = NotFound desc = could not find container \"e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974\": container with ID starting with e733d1f4037a8fd4b99d38795772d12ab26851469ea1b36dd27c56cf69726974 not found: ID does not exist" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.262156 5117 scope.go:117] "RemoveContainer" containerID="3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516" Mar 21 00:34:35 crc kubenswrapper[5117]: E0321 00:34:35.262448 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516\": container with ID starting with 3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516 not found: ID does not exist" containerID="3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.262501 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516"} err="failed to get container status \"3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516\": rpc error: code = NotFound desc = could not find container \"3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516\": container with ID starting with 3ba31250503056e7647c9d16fb94a4f5c52396a7dce9ae3a1b828d15727f8516 not found: ID does not exist" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.262527 5117 scope.go:117] "RemoveContainer" containerID="62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb" Mar 21 00:34:35 crc kubenswrapper[5117]: E0321 00:34:35.263012 5117 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb\": container with ID starting with 62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb not found: ID does not exist" containerID="62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.263060 5117 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb"} err="failed to get container status \"62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb\": rpc error: code = NotFound desc = could not find container \"62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb\": container with ID starting with 62842cc182a7cebfc5f612b75f72bae203e7a7ce4efda5209fdb9f57710ff9cb not found: ID does not exist" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.597007 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bf33857-8eef-4622-8344-21e7a8d0382e" path="/var/lib/kubelet/pods/9bf33857-8eef-4622-8344-21e7a8d0382e/volumes" Mar 21 00:34:35 crc kubenswrapper[5117]: I0321 00:34:35.598424 5117 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a57ed631-177f-4f91-9667-7ad813439767" path="/var/lib/kubelet/pods/a57ed631-177f-4f91-9667-7ad813439767/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515157363745024465 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015157363745017402 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015157360261016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015157360261015463 5ustar corecore