var/home/core/zuul-output/0000755000175000017500000000000015155644222014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015155647473015512 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000112523315155647416020273 0ustar corecoreOikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9GfͅR~i.߷;U/;Yw?.y7W޾n^89|𒆷7̗8zTY\].f}嗷ovϷw_>on3cvX~egQBeH,nWb m/m}*L~AzHev_uαHJ2E$(Ͽ|/+k*z>p R⥑gF)49)(oՈ7_k0m^p9PneQn͂YEeeɹ ^ʙ|ʕ0MۂAraZR}@E1%]˜(O)X(6I;Ff"mcI۫d@FNsdxό?2$&tg*Y%\ߘfDP'F%Ab*d@e˛H,љ:72 2ƴ40tr>PYD'vt'oI¢w}o٬owko%gQ(%t#NL֜ eh&Ƨ,RH 4*,!SD 1Ed_wkxdL3F;/u7Taqu5Ոӄp\2dd$YLYG(#?%U?hB\;ErE& SOZXHBWy|iZ~hal\t2Hgb*t--ߗ|Hp(-J C?>:zR{܃ lM6_Oފ?O1nԝG?ƥF%QV5pDVHwԡ/.2h{qۀK8yUOdssdMvw`21ɻ]/ƛ"@8(PN_,_0;_x+Vy<h\dN9:bġ7 -Pwȹl;M@n̞Qj_P\ Q]GcPN;e7Vtś98m1<:|a+.:a4nՒ,]LF0);I$>ga5"f[B[fhT/ɾg}\Sj#3hm!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtW#:7erԮoQ#% H!PK)~U,jxQV^pΣ@Klb5)%L%7׷v] gv6دϾDD}c6  %T%St{kJ_O{*Z8Y CEO+'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;at 9:I_|AאM'NO;uD,z҄R K&Nh c{A`?2ZҘ[a-0V&2D[d#L6l\Jk}8gf) afs'oIf'mf\>UxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mpھ*V I{a 0Ҟҝ>Ϗ ,ȓw`Ȅ/2Zjǽ}W4D)3N*[kPF =trSE *b9ē7$ M_8.Ç"q ChCMAgSdL0#W+CUu"k"圀̲F9,,&h'ZJz4U\d +( 7EqڏuC+]CEF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}k|]eO,v^ȹ [=zX[tꆯI7c<ۃ'BdIc*Qqk&60XdGY!D ' @{!b4ִ s Exb 5dKߤKߒ'&YILұ4q6y{&G`%$8Tt ȥ#5vGVO2Қ;m#NS8}d0Q?zLV3\LuOx:,|$;rVauNjk-ؘPꐤ`FD'JɻXC&{>.}y7Z,).Y톯h7n%PAUË/_xY~7.w47mnjGgG{9_e552s4IG^ۃn󨔖I@[ tWv Fyw9J֥WmN^<.eܢMρ'JÖŢո%gQ=p2YaI"&ư%# yCùXz!bm5uAߙXC90뼯nNNXYt\oP@gOV ]cӰJ:^q';E=-dZB4']a.QO:#'6RE'E3 */HAYk|z|ءPQgOJӚ:ƞŵ׉5'{#ޢ1c qw zǽ0 2mK:ȔsGdurWMF*֢v|EC#{usSMiI S/jﴍ8wPVC P2EU:F4!ʢlQHZ9E CBU)Y(S8)c yO[E}Lc&ld\{ELO3芷AgX*;RgXGdCgX JgX2*Ъ3:O7ǭ3ږA :}d,ZByXϯ&Ksg3["66hŢFD&iQCFd4%h= z{tKmdߟ9i {A.:Mw~^`X\u6|6rcIF3b9O:j 2IN…D% YCUI}~;XI썋Fqil><UKkZ{iqi :íy˧FR1u)X9 f΁U ~5batx|ELU:T'Tស[G*ݧ ؽZK̡O6rLmȰ (T$ n#b@hpj:˾ojs)M/8`$:) X+ҧSaۥzw}^P1J%+P:Dsƫ%z; +g 0հc0E) 3jƯ?e|miȄޙ-did˥]5]5᪩QJlyIPEQZȰ<'|!p+,ICE^fu `|M3J#BQȌ6DNnCˣ"F$/Qx%m&FK_7P|٢?I-RiAKoQrMI>QQ!'7h,sF\jzP\7:Q\)#s{p'ɂN$r;fVkv߸>6!<̅:xn<# -BȢ1I~ŋ-*|`В~_>ۅm}67X9z=Oa Am]fnޤ{"hd߃Ԉ|tLD3 7'yOc& LFs%B!sRE2K0p\0͙npV)̍F$X8a-bp)5,] Bo|ؖA]Y`-jyL'8>JJ{>źuMp(jL!M7uTźmr(Uxbbqe5rZ HҘ3ڴ(|e@ew>w3C=9k-{p>րd^T@eFZ#WWwYzK uK r؛6V L)auS6=`#(TO֙`mn Lv%7mSU@n_Vۀl9BIcSxlT![`[klzFض˪.l >7l@ΖLl gEj gWUDnr7AG;lU6ieabp៚U|,}S@t1:X _ .xI_7ve Z@7IX/C7@u BGڔE7M/k $q^hڧ};naU%~X!^C5Aw͢.@d!@dU}b? -ʏw |VvlK۴ymkiK% 0OFjT_kPW1mk%?\@R>XCl}b ,8; :.b9m]XaINE`!6uOhUuta^xN@˭d- T5 $4ذ:[a>֋&"_ }Oõϸ~rj uw\h~M il[ 2pCaOok.X0C?~[:^Pr򣏷y@/ڠ --i!M5mjozEƨ||Yt,=d#uЇ  l]չoݴmqV".lCqBѷ /![auPmpnEjus]2{2#b'$?T3{k>h+@]*pp桸]%nĴFԨlu |VXnq#r:kg_Q1,MNi˰ 7#`VCpᇽmpM+tWuk0 q /} 5 ¶]fXEj@5JcU_b@JS`wYmJ gEk2'0/> unKs^C6B WEt7M'#|kf1:X l]ABC {kanW{ 6 g`_w\|8Fjȡstuf%Plx3E#zmxfU S^ 3_`wRY}@ŹBz²?mК/mm}m"Gy4dl\)cb<>O0BďJrDd\TDFMEr~q#i}$ysi*TRJbż31"qrk3S/0g:P{bk?-7B+b*n1[ln 1A 1BH ^iӣ?hl&0Ɠbb__2++oI~!&-[TWvxZ>4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F on~$dƹɥO"dޢt|BpYqc@P`ڄj҆anCѢMU sf`Yɇك]@Rɯ?ٽf? ntպ$ˣ>TDNIG%]TGF[vJ+ Rjv<Ҋ(.GGzpFL`1CS$Ǥ46iWMUF>su0,gy(&TI{ U܋N5 l͖h"褁lm *#n/Q!m b0X3i)\IN˭% Y&cKoG w 9pM^WϋQf7s#bd+SDL ,FZ<1Kx&C!{P|Ռr,* ] O;*X]Eg,5,ouZm8pnglVj!p2֬uT[QyB402|2d5K: `Bcz|Rxxl3{c` 1nhJzQHv?hbºܞz=73qSO0}Dc D]ͺjgw07'㤸z YJ\Hb9Ɖ„2Hi{(2HFE?*w*hy4ޙM^٫wF(p]EwQzr*! 5F XrO7E[!gJ^.a&HߣaaQÝ$_vyz4}0!yܒ栒޹a% Ŋ X!cJ!A\ ?E\R1 q/rJjd A4y4c+bQ̘TT!kw/nb͵FcRG0xeO sw5TV12R7<OG5cjShGg/5TbW > ]~Wޠ9dNiee$V[\[Qp-&u~a+3~;xUFFW>'ǣC~방u)т48ZdH;j a]`bGԹ#qiP(yڤ~dO@wA[Vz/$NW\F?H4kX6)F*1*(eJAaݡ krqB}q^fn 8y7P  GRޠkQn>eqQntq"Occ°NRjg#qSn02DŔw:ؽ 5l)Fa/TTmCԤ{"9b{ywSXE*m#3U ùRIvޏrJ`k|wJKH:O*OKy`( ݢe*{ ua ȻݔhvOkU~OǠI/aǕ-JMX _.6KsjA Qsmd  O#F.Uf28ZAgy>y,d$C?v01q5e.Um>]RLa&r?+@6k&#l)I5_> ` D s5npo}/ؙq #a2V?X~.4O/'|/_|&q̑0dd4>vk 60D _o~[Sw3ckpkpLNa ^j 5*<&}kˢmqvۗj=<Tr=[ a^؃ È(<^=xZb [_tܡ&yЋ{ Sym^?̑sU~' Ԓ f\itu)b>5X -$sٻҦ62L0ډ"ܺ_z9JNȯ=@oUI y4QE[/Y5d{zrBܖ6Hlc "mKv~[uLU4lZ;xEN'oI㤛rP*jC# 6@dmHg1$ʇȠh#CBΤ{sTQ{%w)7@y1K^ ].Y$46[B-3%OONw8d`Q4d$x0t8@t]y1T\YAidtxBG:pɨyeNg4n]M؞ e}Wn6׳i~'ہZ*FU{fXڃP'Hl4 ,ŸqMHDCYZz Qnz܁$Jp04ȴIL΃.0FiO-qy)i_TA|S2G4miBȨHM(2hys|F 94 DNlϒòκ-q|xC ,gKDzHR%t+E/wd#礱ºȄWEz o\JξB.wLKZ39(M +(PWՇfR6#ю3Ȋt ݪbh]MTw䀩S]'qf&)-_G;"1qz퇛0,#yiq$ՁɄ)KٮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?,ڼ"ϴq*(A7? /W= #^ub"6q f+=^OI@߱^F[n4A#bYѤwd)J^Z{*ǥzw73LuaVad=$6)iI gC~.1%YmҪ+2gSt!8iIۛ*JgE7LGoş\bC}O i ycK1YhO6 /g:KT sPv6l+uN|!"VS^΄t*3b\N7dYܞLcn3rnN"͉Ĵt2zG;ro&wU 霤8ƣ8]Z&&t.Φs-6߻t+,8MSG3K:RD-=w7j lW + VJ7!ɺQzi#f>夑3KմԔ萴%|xyr>ķx>{E>Z4Ӥ͋#+hI{hNZt 9`b˝`yB,Ȍ=6Z" 8L O)&On?7\7ix@ D_P"~GijbɠM&HtpR:4Si גt&ngb9%islԃ)Hc`ebw|Ī Zg_0FRYeO:F)O>UD;;MY,2ڨi"R"*R2s@AK/u5,b#u>cY^*xkJ7C~pۊ ~;ɰ@ՙ.rT?m0:;}d8ۈ ݨW>.[Vhi̒;̥_9$W!p.zu~9x۾vC;kN?WƟ+fx3SuKQqxST Ζ2%?T74a{N8;lr`$pZds=3jwlL Eڲ t|*n8[#yN SrA GYb8ZIaʼn8 #fg3i`F#5N 3q_M]j 8E!@1vցP7!|+R@;HspSI]ڻCZUcg5pDcIϹ,oN-_XI,3\j ]ٟ5~' SuipA!C厐$&k7dmhz/#"݃,YqCL$ڲ`"MUbeT>Xuv~4Le͢ }UVM)[A`b}mcE]LCEg=2ȴcmZ?E*-8nhױ1xR2ϫCya` A y!?h!9yL%VLU2gr26A!4vbSG ]ꧧWp/ &ee *w$-`J\ ptǣC^p#_`{ К8EW>*(D{ٛ,[fnY𱹞M=6&$<,"lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPc_|(RaIBKb+{P.T! =ĦiTob d<>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?hXf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O Kޕ$"~ %+qf2ƱwE`4Mkdxv~V,9[V1b&D]O҂:NGIXeOP&o$MQbJ-if~jpv+0.<"*D!E\j,".橴)2r '`RbTꢬTaVܞ%_~CT^8DV ⹜'k)Pbi@NՌ1T |bp[Oyp[I-OxeɠNE2 dA|BSp&'i! 鉎4c `ϫa A=N.s>arA%P%O d|OOIي='nXY__a//;ikie-͗ŕmJeq\ݟpC/xY\Y!Yd/PLÏl%Ba:ڡ% 5帑`7f2@n淅0y\5-ϊj9zvrWkZ?/^w%ò^m.z?#:`0OD=uOV9" OxME&{q6 q.兖oIӒ/H'73"a`/I?_,u}]i$=}Oiض!I..h ah BSkg0tGx:d4]k_j-x*7)wBWܟxl}&ԝKݿ챥{Dh&â0Eൈ1LzӾgcZZ/!o<<^KXoǏtӱtӵ]%DKkP8[J-d1[^XnՅ"wBohP&sI]_gY,i9jDfY~^`R~(c0WO,qڣL2n:l?QqWo ),CkYEGE6Jό}bDuygْae,#)nZa0=6t#7=kOId}n%:=_O&ry#S&[]Idi2- `<"?@{X祼C!@f#>iڊ5Rȳy" LGBx5,.G9/N@|r2]|TK.Fe.:?geK"`.ͪ*/_OrR4O %oIU,қGQ9 wކ?ZxgԴ >P~EE_P5=|* [F^)/;<2)HQKh k #] @dC~'02F?WiFO)ApK "~!Ft,)/E̓2[wYBn=s=\w䨍SxȣeFu;4m#ѻQ.HP,kg(G_ E><*xt,IXlm7ˤۄI}gDmWst,r^T<"ArwtY+xI:\qsQ` Kn(oU(_uVq!`<2 xPG%%Q}}~|q. pWSY] -Yq l( Xb)4@k_BjӡnY'< vA^?%QgFF4"optmu~ko9g 6qEUH>?5L슷U6(`9 O#dYՈ5 㷓߳7_P_B4܊[uWr:jaL}i% S߶>JTq:;1z [Q<]| ˦:W0MB /}!fM8s V 5G-`(8.5*D4!* OzW`Myz0j9EGvv4KcPr/= tY7Gt!yR͎g2'e?Wo>%-B?T [pڸoi*h$}?`ل~u^Bj/>#q {64+RGOzz[N. |$Dʓ?110†SBb^kp/r) zrv[jS4$0wהeKiA8t^ G{3iI]^=u K #Dh ~'M-{ww< FV6%rÈ'S>~HӸX9<HsRKZ NKI )Le,$/pVarIS'XРV,)+58@Uv|r LUV ULhOت vxоY[-)H6J ~9+{I1{RxGBFNC"}=~_Bf,-e[CG ojO W֒ȪgnKk:ʵolI7Lp]H*xZ{̔eD f=ؽLJ3ɛҀ0ȫ5maôMnxf\I0 0MO't8iU>I-^XU ʖ*+)iXR"zC2@ZTϵLѧ6T^`,Ky%e]: 8Tse| U \eXc)(b1=aS̚#X_ N] 0ݦgЈ%qg<A[I0X*0pBrN!&V">6xZ0mE'6ӵvaHxW}uDå V-܂ڲ)U<:oLW\UW)L[Z[D%tjx2!yWZ\fI43%Sв%p܇83(ٺ/`\ƶ u=솗"؄N0ْr [5A;[PM?{cJMֽmP([*kNJ|A]}@eJXW@$o.VW 2aن"Ƒ,+.Mc{Z+dPhlPQg  ")W_$j@Yl܉pslk@C&r_Ռ?kƖ:>tSr c;m I$S7iڦSmW36 "Q M/"K^c_s<=~>^MO2,̈2dM;d4O9::%5xY6+\ 44 XP9Ge}x淍n-M.v\L0lk3nk->sir?kyY`oAEqǰ\ca`YrN\&dwMPR^Yh  +nnVCWeOWzJHaC1$j }D*S3@+BV<:eiri c)pCWeCWp&ELJtU6qݦCv$mЁ%y%6eK#f(2Lg-=$vTmJ]}|TC!]U>zzxu-^*3ttӎW:ҰT:<ʾOa,RomD v@3|U@2 yt,*CLc#'`8ĝo(e壽.@ ÃeFXeAƳM===@:Sh`iU.Qk/WeQNqc,O3U]*ږB0g s!w&t*S6 7r$dC^l&pM]LeqW=l4~չi 6Ud|[/|xUReX=Wψv{l `]ipE*mJ"UGUgI6FHs{_/ Ғ4˫ 3I ՙ֒\f`,&Pg_ +}mҰ9ak Ow7~ՌY1fn`cƞ5gw#~ܼ."f0fА1@/Z 2F-&2Fh/f=~6ʰЌ =lݓ&MSjMaG/|f~L$ތ`6ahB33MߨAfS~AX=P;ʇ9O1콚|XJRcXMHP0ӈ9ٓi̱e sv%Χi/a7}]G]i"cHsuoD00C%DSj6s-g&?30F/"K=ʧMxY?~s|62 X\+Fh}N`1,#0ݕo;a@ۻxy1,f.}G.^E$Q_7*+~_:E Yz;Dвq|{>\"-pī&?e0`Ʀ^ɸD|G~N}OS"TO{ l_M2XRe.I41hs wzRBe>ꪃq\fB B)e.K6QNgkַY$5mb`WIȻ,y'_N޼?IοەKhwŁ(n ܶ|X潨ep+@g~kl |?}?qyi B7ǴVvv9]1S38C ƭ +!_ޡ$A3->@2=7' 2qa(ƚҪI~|T@{pޕco=Ϟ @~7l PGbfjA-"fb3\"؎[`0pQiP#!4#(F(~'`4Zr H]aj45@qxZT2۵`ya2gVZA}kVR"@.w(ܨu> 3| ƺ@8B!bnj<@1;`]$ нX@I^t0ȉպ7- 8Ԧ^ע~c/jqto̳H4P|m-!3Hxaӎ$(u U3ͺ" ` c*@1O,^Gt-]oiɓ@]]Fo@b`y+wMF؇N{v'쒳4_]p>AM[@fB=uR,c-*.CuƏ!_6$> 2I Ц"4HlS4 [~^DҰ.?v4 ksX@2HҺ:;1j)HOra^ߠv#B dyFV/ہ/a rîySEi,Y\I1"dy%%DҐ,\`*RrX{={!ԍ\р?ӁEedxPT#ن .J4ԬaPlK47okS{!~ӄ%_;2bB{S3׿l`{?&VF[C ieԹj"mER`͕p2eU?/PY.#D (Œ"?M TKSQb;Jfty$HbHX#:vivCbgz\zŽ(El .hOۮU聑v= K|\{|Ĥ/AL[Yb !&2C;"I<sEXJ`seL>[W^Cs_T}QkX>[xLɌAQF#=͒y $*K۾UY[a8PV:f?NC"`)fqrdaݮ fٔަ 23̌B^gOd9޽m(XY)p\W[tO͓CQzC{i=Kyoh^q/ zpƎmRe}IQ'􇴗pa=r?f!&>k_LHgK"Y!? ȯ&ѷ BĉPEQ2ɧshT\٨"C zD$ -ȥ.W u;܌ 7ZqVnaյZv|47w !܆˓u%0,.TD|ßbjuNAK˷ꯚ!,,|$3b0`vfCb?r&OS' CĽ8 ! iQ[(` RV9  82`^Qꖌ9Pe&c:_5giƈy11װhw9 x\~+04uWYv/o%`!`'i|IK*pyCٰS<2I6r@z?QDl˱U)Ô9ښ;a8lXV8#i1  L**"RK%r3?9Yr ^i1(+whnX[ihôG60m0(dYE\ n u#`x땺β.6ؽnF(W+^Fu"8>R7Èk[6q8j=;O=b+q$0FNSq=74@ i(p몧=> G0B>wڎ[9[9:tQͮǘ:>6A[ZWME]@="U8;&m ҾfSE*JIIW1[C([%mA(۞P쉄m]R͜'L*qVSm5>ejw'O6W[˷F*"| BUB-'ލP[lA=n:O$قPwPw B uw#}"zz[mO _%߂P{BH*FhDB- W 4ܞp7B'n tI6Y K Djh봸!Twxg4兎fUUiqh"z0pΫD}mM>*s\h;YTO#] [0*,"q|5+}쁖MC_ȷ_3vӞK x8fǪ. K<'oYWyO*8,\VK9G!. ?9jrD=8kxb0u5O&],XXrQ)S,2E2BQ6jB}>忸'ny^e)<)}w,񾤞z K[ũzkrrT,JSDCpuj)< f=c};MbK*•gɜ cPEdhEP$f[8W 9$U@j8OAӯԐG*@gmǏg[Q/ WQdyp}<~Ï=7V;,zuG(r'7a6cc6p e*{Y"ᡷs5Z'?_YoL2备 Y"e\֊x;teln?*M"y`ң70xY]coCa\7 E2y>]Sx[<^{-pcDCԅߕq+8f8'ҟ4 No %An`q" "PkyWzmC%HIOMmx|UTeJ 趗Uc-8ޭA !=XXmL5DEz|gk@ޞ {>T_gJ@58瘧Uc43 $}ȡkS[ǔ K G50lga/-8d_2W{6%ʞD8(х{XF.stw'YB! Ӟu͆Dy1 {.?*ʨѐjREC' Nx}T@Bݗ71"yWƑfxX][Ο_b=G5Cr<5>Ƚѩ(u-3D]Ց#D]ٮok6dEU a?C+$%7ƒ4$D GJbE`zmDjn'k|}3 CH[0Of"IƈjR $G[Ɛ SeIitG 5&+vƿM"#aW@Ot; ! 9z]=cb2ws,9xʒD@-pնp%"T?wLwp=%ݏLjQ-dC=r0Q>&U)Ez-9gu @% n,tu-R-`mh1 =0>G- _vbУw/kpXɦI[," ! ]ejiDxD+;թ#e1Wq16ӴzB-a2lq0lN6}Ia!}1 |vCbR'z=m\&`y3Z,Vdž ]ӵ8ȨM214E=Y6Δ'ZR;IL='mGᕍoa#Z{:gV X @I83CR=wfkF*gm\Hv׻Nju!>kњiZA뗫Xj"j?plWϣ}\k[_[a۰/#*&)'P_7MgU#Ꞟ6-m#7{ɭ_G6s_M3>JX_M$Oo-3Įe:' FoXZk[w!28^YWѦJ m};T-z[k">rX |ċ29-鐮t6UfJdCc|q,3ZjǂKc`'$8S4`3N;*⽼?>j"VWXi ?Q$5}B9\i z 2aN %+!B-<^t[ Jl(b]XxDiAhUrk^7{|Hc&]478*♃fK .TS2p)bI;i6HՓ:汏 !j]h» #BKFN 4YsA3I6+M>hW=d^VHScH3i ;#pYKfT) xIxyM E9{(X դ7xࣳ!pHk:,op|4mw4.]b(m}Ń,!qBa)b=d[Z66Jp&,xiQ$OTz . BbqYGw)Axz_hukuĨ&4,7gMeUSD%]IJ j.@Tʈoc7 ڡݯKg| H~/ Ƴ%C}`CZIANYPp"h)KLt( |CR/=vY / |Hm h_;xJ\Q| ةHgYTM2W1 ;>޶wCnV;P{_T47D. Rٱq ƈKM-MUm r$`Ga]vpd#FQAդ !T1rՊf})άvQ}]5x UUysr]r.X_8ҥv3)p/~G2!\{`[Tw#u}KwvQ3H*Eȹ5<:+5d/E!^o88+&2 90\ XFgD G3.8._ZcU8֟v_'~pZkd/4rGGX$sW]ћB(%,rKq )bguJ+H365X90T , B DLx u[%\Sr0Lޜ5ٵG?ָr[%]I(bMڐI1w\ ܉0ZPċD 3+X!pX;&iZ"iD+jk:R,`*qd5Et;65 i]b^֫҅ ,\7 q{<ܯWu֣> ~K04 883x+INՇ.8=xX%st- #OTdhi G3Υ.8N^ڠܐzw&TL6. , $1+Ʀa9{:eo 5J_LX+Ӥ6;v-j ;ʊݾ NvQ\;%el=vjU圥)U8}޿t2ٴf>=@S7LOr]pZ;+Sƞс<d=Y e6-8`B vAP56B$'t76J× ]AsIg&Thk×6˥ox5.rS?me_tE2w!˵~ $n ܅tϻウe(|];[}j_yZm!ڭ7eO +4FVIS`űhSהJ#@si7: MOOai7YzxX?^6O1>5:.'<N򨒯%cd%3+:h|bs>]{~j}K/_Qg/ENN )z C=}rPsćdMH:W]ԯShj0.rnLUѐ{ [rY&7(U)K삣yaҶ\zш9=nd:zVc4EE}Izطэ5>q/O[Z$Ɛ"~g-tWoo F-gx) mV-ۖk<5^.0N։;3{-36@"QKYNJ8a\m¶=,9!^U4 ; +SDGmm9K j,gJQz lu]rV`e*H 6!c,و֍X&H:Be>OQZ_zûu;hC??4Tɇa7 }tAnpGH!vw}ה:78x)?|"g:G?xR{ʋ|wp~φ*cvlZE6⥮K|{>1P<^JwvF;Di1n౷K7ԿҒݢ}jW=J˞ ^uim0j6>8J+L ꌛ qb,ncoA0lP џLIvqݘ A'(Zb@ц 0PYIg8~ CbG¨l3Ӄ'.uW/@7x`pW?a6*\aj:]Q(b:{#̢ԳPJ`O kT}E37KCa}IMre#w`>cwɪjP'JtqmcrF ^]1Ή! v0C*GӡB ֚\dL 2DZ/5s0(C8@gXKfrqV]rnZ  G7X%&TGUh~tZ3@?=oo'sr4/~n D Mi:@SvC,t֠I<**\r.оb$ح7.~'7zSdt1%KVjy~0S P~)a$_x$?{XvU7ƇW O`>xcgePIuo3PF\&Z(qurKBX)Yvʏ\t+k=}VV-/1F2K|.HYƷ!Bco /^-ksЙrǵf1ʛu_;r{_CQfi\~72-KY|ۍehNžp :QODڒ\ۨ䂽48O hfhj%uMkf  x`›2D"Z?e$e]/o aSs6K`W6~^N!3T^!Nۅi-ڜǑc6Ğdž^—aFB >ۏ  Qrg>i|_}o;p<~$žBb?x!Z_ Qu Z>r2Pl*??z0? &~K1^:OvQxsߟÿ[u{B"Z`0O`#&OB(nwXB2ZR*.kgkPX4mC޶X{VqjVQ*@9%YˬA6JYF 3dMjR8f3MdRݮR q{x ۞@TvY#MvuZ@] o5d "E&+! ̵hՠ :,u ׋HVh0>Sc4Iכ',?$n7XRVGSg_+@b`Q8)=+/폱?]`7+j>E!tߠk`IE -t@Kܦ ָ]-RK~ժ`r>(Z5,jR3YRu?_ QsӋ@g.6#=eQ L=CovKY9~*ûj/e$G#Xwƽ7zٿsxgv`yϨ%ԉTaBǂLn:ZY-:=.k96E\ ?sZ#[ȿVN@;gtDa[jr,jYITSdkz`DiZJapʬVJf' kXңP21-ycݞ,ՈI s/E\*N<?aߨ}@IrԛSR/^N;qцijҔ?KI1+ϨC"Rr+5M촤ͧ~;Ǩ )/uN9 56}aBopiIF|x[=Ikі蠩gxzZD שmS eV. &Z~Q) B<{v;̘6D6[IE Gߋ:(|M6fЁmG|F-Ka~ޏ~x3-1gX\CmƨP$,s'!`=Hz$qLkHm}Pı.!*B[R1Eۖ cͨHmcX!qDm{})彚3ͼVWMK6 +HE]WH5;K"=w 0/ݡUoYD %M_ k^ml[ZmE(U6@MQ:MZ;v302s4xBH1ޢ[=W(-qxBkzxK 0P/páIbNDSՒ35Wyǹ5C޵6=*?.-2()BBxOiͬ$'߈kM2՗ ZJg- X#&bS.+bՄ42zŲ9z^folQq}{!C' kTY^i92׸AYƲe0`8 (xz hdZ=%)&΃'5ʖѴFѴ[D[KuˀAe*3"<A+ S9$B*y8 XcN0 dʖѴFѴgGZQY;0^UJWfk5v=0i_>GnR¸(/_y. "x{j8  7>ȫCvRblbdWFS.IoejهY{b[ gK9uYK`B-h9ģkԋYAa„ښF,`{ࢸ74)s#znrTT[yv;?d R> +fvT;"SyaMpnYx֢nbl bkl^39nh'X@nj.Gf˰0X3h / ?^HU2z#/U@OK :?tb>fx_ ;Nӄg a@oI&k~Povfe.@e &RnҠ-L8P /ibLSK:P>Eh{^Lg<_E]#aَCGCIvf] mXnIPLh`YXVO’r(D@܈`t2SC(HWNO᧓F%I+$ln Y]@jD> KlEd %u:bjgiXbȵeH7K|YZ"(0j_,zڱ3GZ1dmG3b gge\ykͱabl#Bzq) OG_;k\j61I}50?OI]籴<#6Z>Oj`95@U{lj>AD1b#Gf ._x@)!ezuy<~9~i)±D M4v&PM=j˼JTshGHK!̛`L'o7P׋aDKBˬ7kI}~1#QܢSČQ\32E^`HDsk7ڧS $%NGh 5'$RIk,x#PK,S" DekAbF0 е8c)Hgf^|ʸZ1H(pϼy(┅X)U):Y_ipJL O|0PiM!om<Ң4x'gilLjPJe8~8~.jLyNyx ;sr +h"I`9j"Ҹd)⒱VNt؄HbYr[wNr-wl*rI, #ǡ1bDMU(TA9g >هNR^5!RAE,Dl%D$NE*w8dd*]3^*k"˙—V%PߟA$yumhl´"&Kz\,t^MNe5 =G# A :!erPK[r]>ץV wڶl0 fqѓhpc' `}(5^7,hX7 fj8s@3k[D Wլe2U6дy&ϝ2T$01Cͥ%AJIתHjQAf ^,YY`~Eэc 0Rro30.*m(l<1u7ۜ˛Uw}N~@xGZtOtX1&XƎghp1k ipcC\˸놷)}<eˤn@6vv0G]* nxiY`tg"`9]װ9Њc`['+WmB8HX}WE7+s-g'ŧG õxfgF!K+6_\_]^_ݖf͞\ǜI0?|MըDwXGEUh#{ܪ)0߼{!쫆6~XYM7qr~_waA M]&38M^5?3dsm!-~OMuhoXs[D@~xc~"Q,_$y:wE!9@@}( |~fvyVQک-_}Q@+e]4NXHP,$.K_G[ׄ7ZEjLX|1*"`9ȧnbTp2by L/i5}D#[j#Z2v^]k%EYcAhfv~|?~8 lwdbg''=k88[%m)I6}fŇ3ǿ>gӑ?ğ'QͿt1WEyz|@3Y҄JzXhz8مWdu 5cЊ!ՠ<4:`Olx%m(Ih~ .0iPVB} AB9gl_H$ C/Ysi'1@,E &i5@+dToLBB}t5ڝeޟ)OgaY8SzĥiHi`jSpIhA(],?2`R @-}h[zOUK#ݟ`GboHplzV&𷠥6/?uEuSN6鼚ٜ*8Ii>9O|[ ?ڊ|fxS뛧EN1`@~|Fn$NBY S)z߸fBzpӥYL Ӱ@>,8LRCGw%ȈRk ,*p]. (e6ͤ-SL´I[ Vm\_*HMx_tJݎYYČ[Fy!TM=ni5țzș$ D?2 Q[r>︩"q; á kwvGn@U|( .᷒TR Ïjˈ$`18mUGO=go1$!/n\@*!:Mi)hٚkOn`1Evşn/SZ41>H=hÝzgy&EH:mud`V~^8cL={,Ćneedl/I"Q4e;݄|8L+4UlՕ$)S`zew@+ r% +~|% ΁ꛯ$q$QIJE^SbrQi:%CҊʇG|^JfeB+ŽhI} 4Q|ۚ"]B^@ u p$+Vۮ3m-x$E=k0nFCkbrbosP%BZ[h]lr!uRU}vZ yLJ6B :ήS:aYaee 2rW<\YJL6ϔbUά̬;n>q4Q[wyt;eМ ڶQ`*#I˭:[pGwP5)eOXi]/5XDи˅ֻE'2E K(0qch}䰄W K 8XάhXMq8tFV1 fHIȖʇG|'80gE}XcAlFC2O0i5y`(63O^W}7B~}>M_ ),*}5O[ ̭en`[+ GG l @84K;Z!-1$I#†sC`U4욧xJ[V=v:o%=cp/ p#Kԗ lڃ(;!*id\c|0WvBsO$Z=8ԁR x[2aФ|Tօا'h"5 P:R" :@+y\AZ"[PC6)W600Ll#vbh_ N5 Ï1~+%Z~+k$_($tA݉DV2I-mj:3dOMGz)(V2I"Q4qCB>t&҆;&q]B %\ZxC#Z$0>yŵjnzcIALgu;^V~8.ϓf++e9+}(;G\{k9=2}e5YvM3`/ɉ9#rM"R]qM fz{d/gs|ՃCYodad+:apZD'¨!62T# CU h%EP9ҭG(E6fU M?X)S{*Ѥ2rʨ|ȇOҁ;1>3SmSTJ]NHh`}>kEIxC3|(PHZ8pX*pTy Z#ۼohQA\*B}-T.V_B g5~Ysy;U i\$w;vt>$-!' ҿ'H+=;9?X1`-7G#v @{zԷ C𭉝޸pgms㻋c6-],To̊ͣm TśTo'n_8ү9K{wKߏus?{vS~-n4YV~yz+cf>n~_ zߛ?k|9o2 w -:;`|x܊QqmZ!?_mqA-Yӱzw!<xz;\\]p|㟛/ͳ'M67B;CtFmb[ m.m;i7nU4,*.3C>F썾C|g[?#70۶A E?D^)>*[5S^ea[5n`3W͑Z׸؞.[yAگ{g0:F51-Fm킱gA>(/R%g}?;2YwͧO~;jͻ463d7^}A i_?oЗnN9Շ?P) A#@X!@"fasJ֣+HGnRt1Y-\WG|P\\e~ܒU6\ î\scUOw]K7qF7m7L߽;ߦu\)wL}1El:}i)޶~?]W۟.eZѯנ.__nμ-ƺTvݦH/ÕSwG x7%,CA$ $lr$ r֑7)%`r R%X2{_c$O56S3A)%Fa(Gy|I#esIϟbmw !GMwq~ٌr 5c,0D P3F4<71TmfEQM*kRTCԟ~vRuHU +M wR!%IpLzae0dZЃIBU]/LT__(I T[m^dTRE?-U8|:GV{IZ$j~V{xEM4D V[m^.R)sT_RBz R)TBPMo5gz{S ҨKT=oyI߯ͷt4ݵ^#'}ԜzqUm+mbΙ^{rսm z}t!u$vi>;;/i'}bmhB؞:AXFR6vS- a݇8G~k]bvۏ'ĺm zV(tc+ 6=z?4] Iu)zKtzʼn'T'酭b &̎C*ka~B_VGޟ_\4_sEt'/M|vGvtfuOöĥn֢B7DcQb=[i캌_~^WλVɋ||TmK ƻ0D˽p F^Wðc͢\r`` U)}bDс&Qc'170Yc C$[N?jCb3v=6InC575Zjo}] La2  c,lr"PP޴:~v4Q ehN2l9KA#܄CoH 3.SzDv&kv*#H1Kz=3$ocF7f"ݎA /)[40,(4qX 6X=Gyl~c1Gu&!XΑ|bvSOG5uj|ͩyjwipyM 91ʡdm'fϘilNb|8k5}YA;%vAw7@r<! ׼AAZRj0^2o ވGő pDg^zCzC0XBXqr۱j ͓\rl)eTF!n$.x7\!fCf5ɘI2hS'$ǯ\|>x}jUXo+! b'M ^l[312`;Akl94w6t'E @ ]0v<ӟvuWS)8VMFV|rLPWWA]WJ!:-㐈lYA!9"3v=aʺ8ǽQ` ^+rp6X5]9Aa PY2Òcp\Y.K= DȆ2 ldd.F1$IdJIқ> $^qŋ/'5 ׿VInY9i:tRl]H90".RtZB]eUڂQe=cć`<F*+%R d.ߺVm|)hkeSmq!uyk.H9S&\\>9uw%+@<=pj G9SFbu g >6:؁LܵHfH,)WJ%L1e..!sz]yhHĭəRBY;=3bkpx$sxGhm'4j鰅G'FAlyvpsܩYہ;(/'LvN|b`EmxAC7wu:(g:h>U ]*A7R⑃˒lrdl:r#XtxEfrEi]`3 R _:31ufvν+Ò?6Ty R;L5۰u MT}zgFG`<#ѠW?eZ#TJ+ iNpLT|05Lf<`q@s,Nfb;X^b|rءY}áhu84g8t^/6*DdzH92R:E;=E ]:N>Ҋϵw܎MyT!Yx3 k6*TFQBCyᏥO:EU~ UL^ަ*U,]3.r88/$'F9ZH~26Vd<ϴl5 L#02g1k6H1rk"HMĖj"²=0/2a #Ws1@@Ӷb/E|?i_շ/rDҹl#&o}fígT6jx05[)[<="QS.xn:*H:91AHj}v*gY{Ό 'M}؋"lr=.E:[XɅ,Nz)b.K=ZPs 2U.:!RW*[LJN>Q MSFR0fy 堊 Xpc͓xMicfZ34@V~uH މH$8Fb6i^Z'%EX0`82 da}jl'1ȁ>,ö}{7iΐ)QTWW}uwM5VU• HLiJS E ,d5 |BFǡ9 uhIЕs!z-55e"tFơ);td ;džo,1x^&1`l|@'&m#T+B)^NrYrU8G?oSRZ7ɑ+s82\)4A0Mɻ9Wrqc` 5;mEp(h;~ Hݏu lmQ D`ksq s/繠~l/vaToܿe{aG?ZncM.P@p(koz: q`*Q2z{F>>>]w{>3L7fm9-N;r~J۹fR%, i +Ow[E EVѷG7Cr0)$mg+ Jϻ=Rčs7+,X;Ɠ8{rKik=keӏ>cpflZ/=kk?4O7$:^`T7{ݏ&M$Fϭ~T\ri%j|V$'ZdfLѺ*GF8\cZ Tvr>L0A﷨V@ alcC¸7ay ts^b;yZĬo{a <d Δ܄=ް|hsSevdGr&\18^- +N5eL{ 6 sM{W= 2Y(:#@O:'W4gp5gpU8+ ,P43 YnR78jrn6k#xklhf)j 5v} %ӔajW" &xgÊ^7;@ڕU,BO3ChzڊT!xCchl\a$&ԘRM(GE2$\zƽD;. vi=Ke뷧ipE{% ͓U: 7 (8T3(m "d%gC- Hs|kַ<>NHva艤jㄑJ3(8 9T)n#'K.*_} R1B1G*)y GH"rF[ҊpvY V8x8GԛA,|jr10 <:I`Grt s|5s\ٝGKW9K\Oӈ)PݏأTv>أ#mH=fA +, "U"3}Z:al6&iI:K"0r#XjgoS9nxlcce/A!4׏7} TR.v}wۛ#+~(R ƃb}ɎP~%#}Iwb}lt97;Z F_C*g3nM^sZ'(/<YS1l4+jLG1X )/@E8#]{FZeO,~u՚L>2pW9L+au@xGj1u4&&E=0798 -$<1u;6ׇI[ĔlܶۘL4~~^X΋dFNB w (>/o-FHAr Ջi &-ABç \54@_Sy-1n^O`iݼ KA栔< ʇL >Y^[kH%)J) Nskr㵳 i%N0UsQD vRS/Ctk` YjWG3lÇm4֋ҼUb0mr =s1NћNiTј9Lz&i~Y+ΙpEʅ6<!* JǹDŽ2Y@0I!?xxZ="xKh9iW% P¾?5&z>Fq]x0η#  {UZS\>ZRk~)i1Sz+})i=quU\`{R/&LT4~rdz]6-il{{ M1LqSY1X?d9Jmi|hFfЁo+ߐ14JxLW_D.>..3+ *h4&HdKO\,7V+TbJ0XKvYO\NG)[Sʽ5q@wۯEj!ӳVM`>'d3&kwM"Mmas.J߾ QhS^yLjz5+f&Cea {UT辟5_ۏm{ڶmo}5fw0M'\]gEmZK歁Rjiq# 헷oߖF뤿 le߀Qc9S*ntWgQP'=[kt4?v {;}cM-osCӻ('Z7k[(a.xYnI'+J4Ulvʃ7 ̚KϚ䬳ǭw7)NG*}KUgX#Ab5|@:(rh_WrC-[a~V_C7Ī9Hc01"ss%s a*(k(%\pHE(8[P epI)\7n\D`i&|f/~+n@ ݾ2#H0"qN3B.An&Q IEEαAk01V\R +OӹĠBl,8ԇp¥sG$Pa$r}߁j1':1ڬLK1r`HØE2зr,s̙wl(liQLܧ{=X(QH k͝И1 ti0 M湐9[*5eu nn&}Tl)έ 2p1̑=9́Iq/3 ZOiQ0w|[%1scnܒ\afE^`scN"h 4#^M{Ioca }'XC1]lE"jR$ @MLd`!ڲV9 ; $iVa^㢥RŐFa :d$z4_RdԚnuG͠ ۀ8?=ɂNU~T}%*y&UbΊj|P%g]VANAS2;BUD)eI-f`I%>b=jѣP.) 8΁6D>-C`"rjI7۠k@)@f D#`JɃV @ns1BscAXtt kT;$]` 6 F^7nZ[2 IDxP<$<Ʈn6 s$.hȂJ!2@?A. j!oqVhZ|,0$1!ddD(!cȡh-ؚOdK 4]+y-iPF+e#o, {K1MZ~VDEK*,e0$0 ت[x5Cb.+hy6siYx/O״m\$ZT42hfU::lѣй93QdQfq׶YcV5Rq%JV ]"ȘrjӰ[fU(x  @9-1j=#F@="vJTEw AH` H Q9$TiO$)x;lc1l9l 'V")UJ:*Pv 9M%3 aha21 Xehڤj,tMz!ׁ,W (]ڈr1xAT S{[N댞 :h+U|&*3^F@&CP Z؄PY;JA?!]ݲNKÎAQ+MQP<|ś-"Aj+F8F D +@(|8h+E8EHFH ΀z"reT^H O8%7aBF;5>T2'ȃjOVSr8쵤T*It@ n| .zKdAc6FJ å":brʨIxY 6iUA$w.H,}袀+94rϐʏ]gF@ޙB## `j?ozD跳6z*07DJ\)JP<`ӌ6 ýC *xj"Amƥ~{\֔b=n2f!Kl68̢[4̠gg0l+t>9.&Η-?y@Z#ŒTדu^_QͲĨiLF %cږ2XoJ-QѨyuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQu9dcAwQh-`oۍ:F)ԍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨZ:R"u10\ba!vϳ}GJt7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tΫ1|:[Tx3 8~ok-x`^i_?0ON>[jr91 2lUZoG @=mz%`|Gj X(M̲֯6ד?@Ǔ0mdQ7@֟\˛;:c(|/^p?<7Bm}h8CkuQy?;ZT`c2FDeڌʼnJ҃I'dtwpcvIg c:HMkf=,hvw~`5Ģ&0㑀:FV XmP~$`{9J$huF1F;#K8?0.L V #+y6 XFՆK#kb,`کV)%Feb,9u=3ayHJ$.?U1J{#kȏ%f`û6`z''dysu' /& G#irg <ޒcid)τZN#* X=^i'HZghb9߱.Q Ǟ<׎tgQÌTa$`~>@B֌l0'X{&o/o~n/pn?\sY̖9L-T|Xb%;|'uyMY9? '7-aO{,߾<ڠXJ9n:<f_Zzv QYy6 +Qήcß:dxÇ<|}$v=z7|`[V(o{pkC4a e7\+^<]}quǕI6YŸ{0t)>Y{M;6qOw3Oj7 3Un?x l~a_ ܡ|a>_,; $/oE4?+<ćbz#RQjRґ.Fҥۡ*v2;E2x̝(|f4͏vYnN(A6WԒnE jUT(!y&-еEh'UǔE+i׿;7S7VtQM.b|m  {7hogyAOҚZ2oi ;7&&;5֊96d)T2FTp%&u+2&[XD%6 b= e|wC. ؋E TɣB-R27PR"i(v"hp润3*Z.f;6L 䁈:Cu V(VWC+pói-]Fpbearyhp9#J")يWnwO6Sg X&kʰ 6@Q;PfWĢ0qQcCe`:1 BPʉBWj!]td1=MVU#YP[Zzk#*"qP,e"EKؔM\Goj L!zs e_ w?ꦗk+u<ݦ\ǘ2b5x^#Yxl)`ڸ2}mRYfqՙ\ k1i3jFq5R 2j31.!|EmF٤#vâD{ ֵؓJC6G=T3ʍڛ .*  .FR-bU.%ba.Z*cQaM:SjŋP.Xv餀k5c@4bM(]*?cz+3"BI>zW wG tQٰ5ugHPjgHALbVLOG6P^Z+wfgDP3\ݾov8{K?`Z>6Ot7#ߝ;]~~8x pA?=^wl7W/_jW %wZm$n׾Ⓚ׽n?~oxOusfNx/NROVw?ͻ|ai*t."_e2{$b vqI&>| |Gn,sl,o9쑣B5,s?ՎM"0b&kŒs$bs$bc2{IF%&,)RaHI2G 7Kz/6~I#s~I5,bm}XG=8gIiNNCo?z^=4~}>=gb\߼lzLc8uG ʟcOi+B4KVp!hD)M#6DLD,?IĦ:G G *$buht/H12 Y~iࢷ'#?v,s6䕝Dl1)6HIꤵ 5I[:^,mq$b]~y? ɩHͲRr<|%d0N/VOZ/Cs$bmBUNx{>zx?J0=\Ż0Q?J$ށ//_^cX+/O5}C1_?q//_~/^Kx/_~/^Kx/_~/^Kx/_~/^Kx/_~/^KN@?'PǠy@މӀ:Pp“:P齀:"c^@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@u-CLG7C= -:<uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uy탙@oQ@.Qs A{M@@^@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uD|Ջ$(~y/f,]o"AxA_vwNvz7nW w}'QiD?<|):uX G~:$bM@ ғerZM"^4>>.*̃I)ͲB42 bDm;ȢhTIA)V!N"VڳiX^j'MUŒ#R,) ?i{ N<חmUns}u)2UJx`2Jx\gNG {A]t֧uZ$7}{ Mw>76tlP=ش| q\qn?ݭ}@| *Rt̮NVUr պ{b//E|b N"lH*3ؘKDlJ|b͑b9=q48^F4XLz61%2XJmDӆԧbwӘ?tOzUY+v/{Fnd1/1c|Xx 2}KW˒WWlIen(qwHVu0BSYy V OEL&^ X`ZPv"%~"%OdUHOd4ĔSTSY%7uV vaq1;˘A7Jߒxar yҹ&q2=/lBtI|( w70{ڌylFċTm8J#{UeW* *g~zcc4qq`${aqlj6Z^F\5%Ū_7WU\ au 8#q+ĩ+*}WqEI*ypB* X˓W0\4Z.(WBqR1qB*y/鈫4\OE\hIU0O(C5۪J,cpSV3Ni ?"psuq\E RY#lI2C2ȧ۳}훳)|ݗyӯ庮NWa<_ \=B 6AZ8 υAJ^n {1Og?=P)a E¡&7â!`8CP)O˿֫k烛V)8IUp$.o ʲô/fś|s]3.dF\J7/># [rYXwgQ{>`Fm&djgP&? 'o3?S?rA6w?rMU'%R 5e cqk$26(d) -mSʟ>1Y @}*h3Q#5"tO+Vյݲj]5\rsp߭b_QD@q rT Snþ@!7%no}2NX޾YkJ3 :& ~l8m1^ N齦cx֞ :3v;fn7joA|³v4;.<)XrK# (la>\(ԻsBKBkp5|KjS颲|T%;;e~wCw[?JC"38ś%Xgu}J/w_CW/'K(pq#;TٶSX2+ox7;?Xs |ʭųؗzeGגwHQ o[IA۟O~ŷyt5ߍ?g!R &;[p»r[>.+l =mO׉~gdx\xQo:^?edcǬ&6t6FŔgj~>nOdE9(v\r={t57M3Npi0ZWkU))Ph!ooUw=q0m>R}wZwP{UmOp~ qrQ ucͫ~6t)gq'~B?df~w:Nz7~\Jj# 3dK^H%۽a0XIPtE=ߗqE~)V1DBXa)x6.J,-iIh?(6cyt 3.KW^I֐Ȝ oN\L~=iM7zaJ3o\gѴ*OKսu;SX+M_#k5a}FX_#>jhQ?b=d__G WmmAY{d.[-+?/RϣG\=0ޜs&EndI$"\BM_-Y%q9߾ڹg'r#*%X)fIԜF5恷%*="*3ia125x愔wFkDoKTM:#i9aZZdR3ȀAU&P' `9MkoE]1aQf2\nyN(B0L[t,\`.pjHWtiWkx[ ӝ(3;ΠOQ7^>0;B8qG2>HK)*-+2紮9% XNi{EaX#-E E#~px[ JhFY /C9Њ(7aqZYguT kE֌Vuf$+3n@{ R%ǴP (PBN4=}HF<$qr*0 4_ e#X0 UƄisHޒH[:2V. p Ř\$NhtEԜF5恷%}y1 ((60z+H7i]g.'ݱݕzbѴÊd ' 6ӶFRK}6ՇްlTQc(H  Y%U!Hɲ9jDoIT*V"2I'XۏNߍΦsNbX'39p?gqM􋯲rڮH۳f=[İg 39`C/2F l؃&9UJV,\#cr%xx@0돌ɲ6eL>9dzt2 庐5wy$GlNrzI#.ݘKA]'cu2@WQ0ޡ ܕ=mr}(͏wϧ#qHj#>$B;].):l{lQLze'dҿ7eY( e;FtA r+@́-da>Ө.ÀKw;EUl0vaŰl XA6%D^erg4grg$2sC,|8+|pVjN=NТTu-x^h7F鯲HTQM%J xe -u y7V4˦,rMNhi(9!{4@90]I(c/˻VcxAt̰̜iYF(DLw81}TJ}Rޖ,! 2DK( AT(p4L洮[Y1q4Ξn.YG~MrܩOxO=N#d/%MwZ0 X:".(GNp$ցOf;?xnoKju4묌-=NC'$UL 3@N`bZ\NӵDx[JqCޱ~@ 8u{RuYcYlQR,BhDo1 9õO£ ĕTh[Y-KX@¬AKVpmO=1"ؑTO,SC-U9+GU>GzGT1bHt-F.\u. ZxL&A c_͙nsf^%xDg?˅_psaDuyaeT7ڡk`OӬs K=>5}J܎J01%j:ã9+kg#RSA,VD)K6*^D< Q<%9OA(VxA: )^ظm.t u< -ɩ883QEJgLG8o]<k2{xތ yZa7h^! ƈR=}T2pked䟮f 2ӺP#J ^G䀳cGeUBRzBRx[rtc 3" z '(q ԍȑjUV-Y-iH%mxVzA@vgο Iq Fv|n>uc*1 . d Qj E-U5u=ޚxëdK=1JOu/"t}DqvoSybO]ER^VQGzXY٫Y;[̻N(n6N&oVvwQs,;g (k/i=c_x]*FJob9' 1h+.R[~"w$TF#8hI 3A7C{l u* %)+egE5$AakARм{2!( ~Gee9 sY-I:T Ƿ"G+ɬa:%F䊘bF(m"wFoK͝}de ų䬣G aS2#g.*dWWȨwNE.Έ/ rr.GB(QW~tˮz| sMoz\A9zezGO噹ĬC#;W[r=${JuX+ sJ?FuRg$u3wa8Zds puQ㭮hZ6 -u;T{-:uྡྷ=}PԯJ߯^ąJP{ g&,) 3{~Na43tOP؋x:=|_F\7g}:37S ]'A(H?ILc@~nM: Z8"x U_La'IhA/%M:q~6|8hfMl7x?v8>ᾗ=< ᲟNg'v4-|\?d&N"?xMpt 6p[W@gw|4I({m$иwpZ ,vId\ (XrGJ$[dGvD.UQUdGщTk=Z\aeQWp.&]pA]̩~Dfl#^џlX JkLXSg]?;c'!SBeBqFb#hdJ<)Id%5e\O9wépt7͂la3if'hZAa"wb|GӃ jNmiXAlCM 綂Mm\jA |Ulx\XW-~G5OO6j^enmղ{mi=Ez>zm_vWt,uP#}fCtt쾠˞Ӣ|ϕ1v%٬ٳ,qKNwϯQޒ 8گjZ_ `Ay,;ܵIrIfjzv@I{LB)^?b5׉Ag9o.{#<4˺RA"O9IRBnES' VDik/8_.v9^E긴Y&VАQՂ <2@*ꯘ6bW]HCFmE|VQ1kr~23PQBRHśՇ5tN_Z`Z +'[|%ZAZ5 PVøD9ضVz_sA o-9po54*r@]LJq]1rWʳ|P h8{Wh8LoE;^W`7=6+HQ"~SW]u:pCfi nOkTZp I|k\7V<;* =*s{%#H:4 ŁK-&¾ /0yri煻 o6;7#% ۿ\*IL7/Xf, ç [{7ɷܘz@2S+Z]PywA=u2Ń>,00Ê*Ϭ- B_EJ؀<7`O:F d}.uԉ\,+03*sZT[=!}ҐqE,JQ?:v# :5״ED_ݏ9VpȥYb=y[DԺN΃h-:`\M)D`v !vИpf #w;ҕ}Rҕ,$ݡ5}=[ŕBz&*H=HS~!)נh%m˜ *Fk̨AIMӰQpsf@:3-0A! Si,9@Km~)SU8}"}&9D//ŀcE4шd%ktlU0ڙ:^niŲ8Ų@\X̞Vu1(t>1bN+:ض|99wDXc!P :픶piȩc"W'0Aʂdq2 Cc¹aB(Y_@ՠzhL"#Ûʁ6Rۦ |ӄ}lVTk /!ӃY?17 -2  A3.s3;B^kq~We 9+G TLEͭ 8g*!ʴP@UY۹% G j!,_W5޾}!/'/B&GY<>|B!{hLX8RjRheΡ t&|% >SY<Ԟ"Cӻ⒄unUydLxMex,!op5IG,pd} ꉃlSYX=hϒ7OYU+Y|^VmqA# NpQ8Gt;T-CŸ=4u{!Rsы0t{l4vŲ8Db)K-Zp BmCve4 q$3QQZV9u104ZgX$Y6 AI5N5:VdAG'xY(刍 ųveԗ+8HEDJͣDcH: v; ~)D*p:rL.|*7Y {5E+Mm/l_$ILh=jwD,Qݭs oeﵵ,Rr>˧A 6И0U2|6 oܿ p@v=4&|n0fϫ!-Ɣg9W~ߵ2eGd.:0)u.$ߚ:"¬_ҖFr6Ac'qJ0.3sY**P6r!8.˷4=vR+ꡐbQU=v,(ڎ%Bnfy(/|1Vъp.xNC`p#CgD|[ޛ"H쭓"(Dr쵴@(1]yWvԮ8S |zڑJ붃*I 24   OuUA{s|%y+;av́\E~^6'=Wd]$JBO6' elB\D)MqJc,bb͹{-<p9I$ֽ l7.^$RQ:LTgڪVw_$BQ4N#5KSQW7N 2L„i#;:Cuf;DkQ PY :ȣDCc I!)!Gpj_\~KY4zEJU$*GORfB(2Χ $pL8"]Kz@ rב(ZJ7<~X}@£aD6y^iRW[Be dڀGF$=|7f[0[ Bpڲ-WZC4jjׯWzkO ƿt۽8F?1Vڵ|°%J-J{pbnEș?a8V*>8#,$i80~xzĕe@"w˹1j>!(+L�Z/tQ R$ʪZwD8ԙVaa !S\m{Y<#'}]dY>kHM: 3k [ HKg&A^J;9_ \I~ܓUJٶYs녩3sS<뻛__2.ﶌw^ *GֹP4iл| (qCTX18R80m&,<0m"Cտ=(دmv!FB{#0xE, E.n?"6Ě"TGOVYkE0TVø }UTgC/OR;,hO Gs@W{ k!($|]IZwdzݣm%ǀKz+ C`9+24Y)~sM~ ]lr'Q_L\v,\ߛRKy{>d(e ,7}c6 b7\kѾ_?^8ZsS؅^;SR bEX$ҙS\o)>M5nϴ繱:/aW]4{rGފ0i7݂֋}*_.\嫛-'oPPϬVe~l.x_=*~Q/[%`Ul)6Ez߿ S M`aoM6zٷ_o7f]ՕzU.z%Onms=#>zCS -N(2a gaIPš"!Mu^5;[ }"kHVǣ[m -a\#zKT9q_w3^1CqRs M qOf4lrW.s]3Pnčt*+::,3 ~Z( b+'Z)b uwkfpΈ0G@5$6)%ܣq83{5m=Z0R<N=$!զcs >oth^JiHD8u$ EiI)%{FrPtoLW0Ԉ@ 21UyQ?x$jʤULL"\z4(45,לA{m,,Q54I.};kli)R'ܶĂu!%*R-ϡi 5״w^^A@5k|B1\~@lJjk'FL#=e?4[%ZAr\0T0!QE') ٝQ2fCOzh\[Uv+lfiJ$qDڝ )a}DH3"+`͓Y߭S*K(W~p?-t!i_Jw Pg_LR"p̎Gt3ԍ58tQ%ߩeV<Ew\2t-ךtxcNHjY_Qeܲ$ 0KUTUz\C?ܨ~DfSۓ5cozIٍO@{Gם B!@,1B 0QN4/Lh4өKU"{|M3'17x.ifsxdg 靈,UQ>W:3embl!νYrϸ\kn`U_Ϗ]3)bLZ|U>{?uv|3U춛53'T(ͫ(u& * [hʚ;osje.~sCѻw23pΊ7Ͱw{F;,5;G( /e~w{P~,uue|]zm޹U23_-A5{WVf9U:[2f5+3ƕ|̚؅s&oX&_9Mxk)K1vu/|j%Ȫּż-| o' iWaYFIgTuk%a ͦ~߭7ru|uT.M5nj52A}B{O=&P0pjTZ*zapt|";u/=;Mof*SqCVHIͬ E#N[`gпwڰ^|Er˩&F+RpW gm?Zw#Ts4kZSA,o+>kb^0>r2-^ILZM\ w$qհ5`X \x\'O7f?ȗV} c7\*(Đ51wX gwٍrQr@ˀMoOL4jcptXp<+pvE!+PDKܣ=W;Ł9HC8ZrӚiTpn~ sVA׾f׌zΪa\s&qPLJb aZŠhJrYJ>tL0;CQI 5eE#IcT d0hƛNjyK?gh4zFp;ܦǕADN$`%*-HVQT$V{oZ{6_kQzXy{ix]b} 3*qq5g1/v8 NxVW$:C}bUkR4D 0bCh6cf0I-P.0qx7 S|n h,mCm/1F 5qvN-l n |?▬ͷL"rWo.&((ܦq;l!RTn8"0Xh?!;n$AoIrz5[$%g\CW%4o~Xg:'FYuv*LD!Pt5<Pn! G^B#8 -HHU|F|F:QY,e&\EH%A]EyV"&"ܬ9[5e$9lǚDۢV k"'LMw{;?Oqq۳ {~<=g\`NE?[`s{3樀XǴ/AL$IiDh­ ];)6ݫU`n%y#IޚwԎhCO}Nݢ>[Vz4b "|ժ$!U`߬ ƻCt!:H(1yFٹ Vk, wQ]ҏ_W>#x嫢P:ϱuvp EUB+^;?CK|Z"h D-N<1_k$cIEB4I)é|Sॄ!=Za9~g[:b H7+m0nﳫ^quM_/U$ –][; /,3viA<ҞH?a^q mGWu^-0bGqVqLl!&oڒ{}/'z* ek,d679lgyM4[We*{Ȯ0=CvM0 (/˛"81HEܤ)Otv"" 5fvFK( wgj:pr^r^z^(pqćys"%6Mrn,`2Km-];9Ia2m$Vs 4`/K7!< 9ucOCR5:ٛm˲nӝd/x[3ndY_vúZp_WU#&zuP,i{ϛ垝n|^p¹_y$UX,)be9U_$nR  GCVN$fH]ozݩ|kTø"B1%9Y, <0{G1@pBM$0.]KjzH&4'ޛ ,zm^ 0ziøҌPL!'NLh' CޯkZFzDQXVøD0˜"%ZAOאLwd RrJBST=c,[[`R&*x_I>.5h8W`r2nq]ћ!3X؇SMދԀ694%0\2[ -O\>.NC6O4Gl7]۽_JTеo AmajWpts&!'Q6sEVL DeU&x֦ _[|75eN3?av8N$UNXtp)4 f&ڜoc5k |8޼xNE4aE@ktSaZ5k2\C54XsOVe4COz%z:16|P3eZrsv{ba<| L ,aL}Ar]8(z9& gQ >_SByzm^9S`%|3_%IVX}-^kxF9pT4.ɜCvxK쥛֗)cXD}<Ϭ쒻o>_Ўvmm0\9@_#ۥZ Po@E[Gu^^j|qq :`boKXzPsV1wh^P[/AvX3nC嫴Tɼy7o#Mzf?P>N g- ꇕ+gY@,Kqa~||rx N뼢ws@'kR8 lC{?(SyμEj.WYn,Yp;fp+b3`)nί mcQ4\$ߟ˗| ^}3/?lw~>:]+X-1)QU-W400nSUc@C Q8RTe^T@i|O9iDoE]8sL~$4Tt6d*uN\.zB;Bӽ{I堳ej60ݟ9o'E_9XC]|LzuT3.uf,)_Oqp'5+6Q=8_oW:Kӽ;J#Щ [;l`:{>Enoo|?yuecW襷 zV /5[CA,`gPb0.e$,1Dq`CwTX䳩&iA]!$`4 1#i01ڀm` wC|"PYY)3jL^j{b6.pCq\(M;$#] 4`&%ueKD%XWtp/pO @Fq qƾ\\O^z2l)_ L\#$, MHCړ1^&տ8qqQs;2ό׷Y=Mw\>0MIBbgDZf{+~\j~.+BCO^ͯ'* T9j{#b|LMk3C6tm"赦*kTøa &oX{y gރ-KWJ+鈇u&1ar e"J0)ұ |06U㝥3@5E?2WZuۣ!Ja$ h HL?S℉"23쌆h_@|Nz0N}6u);޿>}Wø}>|E=^a@ݻu;A\6 {=Gf`>G*a yuS%٢l tvɾT1T^\Z͗.!1j.Y2w?cr@kf%13Gl\B`Fg~q} 0=p>PҞ8ǮY~c-C-*RHͦu߶b LǦ!#QV"orsP&oM :b} 3IX_~TDif杼P{m`y,dBۨGvHUme6()o`-3J5d:J;ɜV y M})e $Tb ִ!P,a!HkW*@|̣3sBQE}ldbm^̢?p.o c-Ȱ/EZ^ HYЌ<[7AUֶQy +C8cy5O|,k.W5uaҮ%q-V97423GDDi٢(fupno;ld%X$D2y}zY}b|2DQ|6ծZM, aJ.m &O>fdaZ̞!oj=X[lycRw-~{qղ4c@C{O]#1seIzv̧ٔA,.+Qb(ȵ-{D"K1/\`֨z-cM0q/f=F^h~t(3eL~} K S ZNJ1@y+h0BIuvpPXo'p$~jtq 'z~x;zWQA/VIl y/0d9*ƶn$."Cq-aGFrUIabԛ]GdcdS͓H/"r(OFV0&2%E`iE-Qm)Dm)Me)F^p쭅j]cr6mlA͔Q NN RL 'W^yq؎ڳtITf֥O}hݛ =42SPFZN 1Qk̈́5MGǹ4ɷůK+,y#la!B{[E\آ`Vk4" m|;i d/ %be"wF^XwQy􋇦O=XJ;q[pwpphUW6mӂNV> haTa'Hoo++y#Lz܃ 8`Pm+~Dqp*ٴI6Ht#ZQ{mEZV#vpkX"8- ~u;_M% ڮY9}R_j0Dpe5I]u+Ua&1(f}: fM8AbyJY`u0{`*h'J,=zjpG9mћMĨ蜶2HdT~w]ﻄa,#?Z9wfPoRLRk#FTeaGРqZ.Fp^0)Q{q"T們ۀcTjU̱ƤmLɹ[ !y"rd8H*5,'$ɵR>Ȩ yfw1x4PDXh֛Ry`v**k8Q +K <UQy7rT Ft7sEOG;}qKuD|N&aN @LWبzSj*m9%n_ף 20| U%7f 1Isc6Πrr-G1!r;7w1g)Hp\Q15<Ē1%@ucv03:1o1AD+E%U/qg0_x7 ?I-/fJF0H-\}άa AƧ&#hc|%?m?6cNߗQ~צ-Vw O߬eOzO_~_sF ?U Te_ϳ?LY/u={߿׭VgfϳWn[6t3+*wUSC4+rˁ&QM/y/53/O /?wzJ7"_Ngn^X7rZ eiFRsRנSu/Uۇ  x*C&zưLHd85pQov ^ E?4VC`J8G 3[tQ n%R6ۈHbOK!cj,%y[`A{0]#M 'כm-wQ{hdf4::.dgK^5QA/fIv  5Vds==^隚}[&ek ь;vQBv70k&UCO\AVT(_<9u{7423G^ gB=423'CKQ_63\4#^ZԆ}wRpЁ2]CLTfn9'g4VTpA `Hp8n7uQX[]7@ǔnh 6vvO ~8kfg/OkCs$;1axG{7rpL^9& }%7P8=423'iwEZ̦4 L3JyeUB9qplH`tvBA9~kEs?Yv T7ԙY_g`S I.E>t+`x̑QyMn%Z}D2ǍNw 0Vnpz}1aAog3ɂ@: ݴ ؈"yZ< -tKZ҉G+ v9x 42Gw_tCڄ@3,a"*'mamMzhdfqIv9Ƌfguƣf Y=Zy%X ?"eP(0P_o?FH-6{-$`dfHroPH*9 4&!l QWJ9+TLڕo݀kI5CQ`:<'`Q2sͥ\usq̑4}$>< leJo0~'~D*'tr`YF۬lSLQP.1̩!6PҐ x;C4V2RB<=4pE Sw9RÝ@#ضR4`MԐ#¨70Y6Z]d!% ceUn ,1":\j a1him31h KDIh8Rs!t4b$.+{X&}J4[_=:0r'+}9H1ts3\hZǁx\8Ji?0_n+{|`ˁy$_pOgՉń I<zr\^q`s_Nҿ5f\XǁSrTReayiJxGUJ`I?4~v'8$k\$ʹ (jfD X`$Kw&[Ԯ,]W{Lz祟775\L]{5V}kR!V~!ЅÚ߃qK{QYX1=Ƅ68jPn[3/kw0(go bUp|/\ WXw@~@צA{ⷧpKVAy؎xxT˘[ ܁u?q;]n80t,r'Xm@W) !.7Jm1[~re.WQ'M֍=o9PVHY@O jxޝc Y_<١%lg ])#t?Ҏ絳^|/q5njrÌC-JݰT  e筝y(7]7[WғG8Sm\4Q sR*8tjR:XzO"ə"W+i7ka/$oSHl%!( ,n5af9ŋ?Orc(?짅8J;c|0O5FoJhoqӓZ\$loea:F! HJ5!A0 8P!A?Hpw#Ux wBd޲3D-9.k6R2*cťmD;6yG"h^'`qcN>9rj'?ɧxj)1W\s͕c֚cD7\9nrjs˹rT*ښ'hd+GjrqsЕVc7W\=sťP `9P#W㦘+. c7WJ֢h`ݤv ]9ror7W@e2ckQFdZ6\ 馘+G-=AGec+Aic+hr}QO\iJN 4†kh&)TDK$i~#I-m{b)Ƥ ɍM, p bƴ6 cq5#4%^H~,UӍ&p2\EV+h7XJI"lRآ*4iba)W 77Jr`"01;}^08[$~>}qWP]u/U{s3oy{ݷW _G`W,skoe^s^f ~/z7ᢇ]o])~ĥZvf\Nص܏wk[N}22Ϲ9@-2uͫyKf}]Ҝy^b8 {suPq.`dJa#%J#mYGZ'&1S%#,S#d[>]9 PG\-'hΎ}wz00`&{Nu~!Y(O䊡۹3lAihܻĹE#VBύ!wjߕRwc_t߿L$Y*!?q0e]\rK}H-#tgWbﻰFYeN̒N)n^׵2k ]N&1.Ld =EOFcڍүxi G[(~\)Yoðv܀Ztq1)=!R!ae_ZAH*&UI"=[ Xi?ﰦ (U>M80pqUvCG VA%?vws`ff@AsT尿Ao|9G3~Tݍ]o#3iǼ ~ |Z+\6(s_ծ5rH`{|*Uc{ fonQy޺`ow4.m*8;,piiiІ#A@ƍ9G[kў< 3QQVW7"R|O#"D͍I>]A]ML J9rN)Kc<&(J8%Q$?"D4%mԷgw_Ax_w kƠhPf@W=f@o:- vH?~XgHM@ 0>Z٘81\ Hh˜H',0sD1Tk%ih=e;Շ16>,s ;0HkΕIwlS*%oj'cE9d/«r0Exf&T{4yu4a MI¤MQw )NK)PMbJRðս}LoܔC"xT*t-2oGNE}UgSϦe'%sg2F=;njf;/ć3f( F9S ޿{\#ѡJWdzxd5 g|Vv켵Y[a=wy b&֝ȝb;aS WWl>5WJBy$" #䎿bgso811]dN5neAK߼zQ>BLS!Wf~eɲBMV{HvKU3)W|$p(eߚ8 $#l"$v^(s 78ie!P=D0 bK  Y!IL{-c1uD@W3\ȼ+|%R4M#YhS3m >B ,n() ֐JV"MCq,>B %%/"DD"jgxGj@(AgXE`5FG.0b7Zҁ #A;t5D`B1f 3  ^cP $+-Pp]q8@|S"KyaQ1 Qќz 1J_L`1C`0SK:@(A<ڞo%evInpVMXI9Jo .NbP y/JOʼn\KH`6%;%r0D03K;9L,9 1"JA;\ ~9b 9NpJv 9G4s ^m━ b`,j X]ep|+*JBY!8GcqJ WCH1!g@(AMpZw`DigS\bR=v (%7"8(FΐQ +%ay€op&IMQ@IAb K4)s R:#1q{񩶠GJ/..yPV8pI;L.KI J!%'_]]DI\tMoTڦ0P[]LhT)2 \1]JshegfA* X.f|#ᖂXQKB(Pt@pS&5q3h @;B ,^͠%c(t-D`0H "~B 9w)8^cȀ뉘[p" &TK2DŽ.SO؊o$JXzI[e˩p\:_pV*bL t5%`VSǑԆ!6=wFNCg@(A<6Tt$@ !via-X ~=N| K'ntдHb rPfwX+lpQr ^i\]{W%!#C`Jyn, >B եń氠Jkk;<89 ϰ*C2+8Y_&K">B *;BX[qTL,զ ΫUăgP ~TYY;IVq>PaBuPMDu<`Z%Wx31A,::VhźPxUie "8"0Ee:°'m B9STBbH:pSQ$TSJ1QJ 4'wás 1''B ԥ=eZl ,׌ɑ !( ~@(A]zBU-% s $fCL מRJi*,7#e6.JTD/"!q ˁPx1/}`*49b%P1qނ/L(*&uOr:{VcϺ',{ڪe_N]QL&adUBwhǮKFYSW/P]iJҴQW ]e[%iKTWW ,g@6uG &,{kSWSWW*S5tj/j?t3W-Zϣ6e=ԕjꩢ'\h2%ut2 5OE]%l?vuhRW a']TU–cWW K&z jba'TNE]%lVhQJ>$!OK= *,<.VJ .\3)dq+54ZA[S\aJq: `2uTJϸ\Wqc*PtD׾]wyW1>6 t¬:kvfE-Sp3 Y> zi'" CVѐ4ՙ=@d!?S~Yx!zMIwT[b܃>HG+hWjjM~,~_8ok8_4T. rB틀0i_L3EVo_,uoFk_|9LEx➧n&15 *ŀ= yXqG|_fIf,A1ԟc;1(7@ 3ƍwpHOX9_|7MN⾽.)̷0S`k|G`z3_9*p!-eF::#q<֡Z3.N~$=M*)Vϙ%E mR|vDm MiQ4,*+߁=iHUQM7xqǿBƻ~ScH~~H8M?[2>0iY}{S=,=IHxn3/NV;Mhm*Ӵ @OSVX?͒ߋ|ƴ=q#e{~T35&b`O&1Z>3D emMXƴŇ1->)G]ﳚ!IKCұ7Kw,܁=iZ>f>f%uJPK ރ>i `Z IKZЦ bèK)_ucxT=gpq[VgݯLfI3FNQ>ܱFw&SDr4וQp9tǺ7A*\7iE`?Y1_#%A~'zJqUo M[ yCL}ilo1\J}BpyAQ?k3A)H~;?Y]xch{0yc|Oŗc|\qk:p!oL':wY.K:22lmU^}FWZ0:G+f*Jn3ԡSךXߛ_m*V⪉Vk崴alM̪xfB'Z',<}X a&KJ1>-Ǿ+a>+-J,fC+Mtōµ6"2aW[Vdeȑ yYAU`O4 )R zT4^L r$If%&z+9ftvO=|.OjK`9~FrH֮ ^x9="߻VN 헹bq~/:Y {TW?VOπ h?/ޢ??W޶,!h<ş{s8<aU[/:'ygԹ$iM l3ߝ @\23{]!K9͞t%D՛N/πk\ ۿçy oS _{_?(TpASM̒}PH"K}@\)u~_6 4Q*4E>u>l?`,b3;۷9H,gଦCwٮ+߿/x?xp+z!%f?@ޫi}{ _.SӪ#p, j|_8 JXè>Gƥ'&O|& +gɖ8_ޙ,F+i ;Fه /N6ErHƱYMd v($;.F `$ň/lorCLjatO??K.t+?}]ȍW+G5tM\麚/lwBL"2WXjdRH+\o>ǃ+~|rco4 Q4(7o:"l=yECooB~eܾQ@a^C7!`)U!IfpK7C+'~tHBw5>290́e50>QRY I(58DLq!1`"9qG6ڱOӉ>M'ߟ; !E.2/s.2/*whMhyI)D2$8 AGS4)IʡM'jM'lDlѣ>1@6ozI^Ŝ j^g\\qefX_/8L/o/M9 *9#!uruT6Aj%AgYӇH[{,rȹǢ{ٺFH`7/d0NVH]]2$-ie//lh6 h{>Ŷ~p2sBڂ$@OTǯL>* I!d?C|D}NK~ujWRhDnKTg,Xq0A+)5P@ <M/4m2ri~pctr_徤D`9- Nvp0p'AKCEQGgz .ųvXg.-#y|AYzy>!C6!)W-.]d=.s2g=.s㲚z1D-7xTr5O`9|(,U FM'jTntZ'\,*F7* %51"BD ,2Q-@d3*xs1J.])JI Tq?'[Wgc(sA0D͵WsUqt9=nQL%J:iU?gx/Q!29ࣴ!w6JA6K*CǫifG \R_7/G_Rr.f7o\GDdĿ_GbL/FOonamm]a.yvβXpvt1Z|- L}?] m8~:{P0&;O-ĥ+ cV-6fzy٠&(q]:%I#mQH1ƦQb)RkT)@뫍 kL1"\$#FmLgv*Юi!.:;hrA/^"R I.uHL2s:BQx8͛OnsWH'DB}]lT՟:MXr;h)  +-slv/fJ /:UeƜ{z[&oF;9>%+}U@z NJRA)́P`kD_+:z}:se=b~d]3f^?o,2;M,+达2Zv )뵳R|"ɹ/ l"Yg2y<|[kRNVt>BZZ+zoW$.*c9{׶VTm!ݷ^lWlPZ"DZcᛅVqW}Zb; Wr˙:YǢһs#ɚ= WOh?"0Teh'S3Rd&ÖB4iI>&pAhƵ\+3Fp(ϟ!zv!G+uo|5VJNS-͛nWgd&YQS֞g\w+') 鍒%A`BTj̺-￰RP9.; mYUެ=ۻOzxf3=]rYۏsKXl_ rU~z=@7ޣѧ6&z4\h<S6R8TzU+iՍitK)LG^CD=Uqp[#啤֒e+G; )X[OhcyTFG.aFLAlBV+h[`Z#`ǝ׺ژL(1DE"*p,Ն[_u!°[ƀl݇,C]ܛspmH#gQDCxIDy z%en)E.G-\+ qPstaFYADsPZ*F3%DGh}J Q{q,嗭ZڸCWwVT6Ec[IvҞʿ/@2%Q2)~ᓰ҂J @N0H7޺V%ADοAa7 Bᓐb9E0w`q> b$)*ƌJ1xLE!ļ1'zvUĬqUԧo=k7? |Njg[ϋȇˏLk)?9Mx\1P4қʊyN pH歖%&;F @Z?LJ ry|̋}ч?bϓ '/U.g_N.mzuB.M9>7$rtb-LS[~MgL3!!6>~b|;]Ke w9AF5.Px-dSyɊnz&^kE{O)˓TʡRӌRAJ *5>1>3DPx>Rg%1K1Y$ގ' 5G9ø*.5\DI/]J"{m)(+EY Sj奙m\*#t3~PW GNqq{>MrrSMO8# a䓜t jBץB`@P{EUc ry4DCŐ`W*FBׇy66g!-92[K׽,3m!)tjz򊗈IoSs\Qi%v̓W2( PJ`MIu, ; !' ;L0e,X yi"MQ!HpQo;cH8f*% `F l~pƂOL5_ds C@zܠ۞Ij#.Ya !ol n ~@)rL^-~ܓeON/PL=9=QHϻ' =9)TjݓÎ{rOðke9)[DIJgR%%VQ-:y-̵ Fބ$haZ?7/һSY]u`* +*3J"-Jd.'3,UꢎZBQM˞x3{^Bur0L2.5cC1ƭ9p1%ۨPohNl~uygkl.d#}"t /iZ1g|CC9ȁ|&ݕ Jxipˋ]c^(I𱗨A\#YwQ"4/ŹF,햪mdmn˿kcwۂ(4i%0x].h 63vݎ #+c>-&]9 = [yOֵ;-;2:'JȈR \%p[r+{l݅YɲY;3I[}Η\nxɣu?$ZC@ꢦi=]4V' Rػ<"Fڅ]0F[-)Hnk;N'@4[W !^_e\EG !o 1F^Ovzf"ϼge]hyqL-)jdz^#Tqj /X9~Ȁ@G}\/LaJ8F,GX=Ȓ/hG_-V0 97,9 `R|Ǚ>"X dM0n D&B wVs%NPV;IUBe Ѱ ]1+"d}gd?OO?鱙NO9t΁&F^{܁',d9u}T W>o_&wK,plj ͙Ą%ƟH.V/>!μcrPc =Kܴƿu1iywYC8^OsMW?O*Gƿv~JOjح%TYQiG@ KYX夓PZΫbKk+, )k'ez-O`S[/0ڼGLr &*ͬqSB3Pat{1{O4F}xzPNσ9b>oώi"$%dŨEx Y=#pKukx6=gH "])eZ y N[R:A^0N D "xŎ"BU%A` ZJ9ч|le7>գ hNza-Zc%uͰRÈW[\!R(gAڒl\[6v75ɮ]M7EpXψ!,~I64Ln.!?*?}MD\0.IozM`َ"4sޢc3E)9 /`cRT#&WS3TGzq7=GbW.p=JY~9H#:#qHpK ]63o]ʠb@ZUcx@Ɏ)CFgT^7Cf0&ŧ 1㍃tGcr8B/&7#x݋bGP:z짾YUG91PpYt̊;^\y="&>ϛz4NHƢ l-~fʢzt9;ו L VB%7{=]{뗅P5Λ &^x0pf_ zE*6Ͼx4yxe PΛ{Ǖ*ht>#M~]tۛŌPSV6^H D!ӔȴscGPBȎRPx9b[R#1Bʾ*&0܈{џdaoÌqzB22jl]o˅n@Ё |ߏuF*U?oNK(ȅbposQ(~#BJRcSgʅBBLQACҦ]8DJK帞==osPH<AM41R *"I ZZQ]p V(h9yR`P;K9#C{9+p cHҷ`64u!!D' Wh7$!/'Q͌}NT(xH*7E' ,mLˆWWE 7rILJײ%v2qJwEOO]HG)a]o?KGψ"RM) N >_5SBj勋I8G"df4 cL˛QhlE`b:}x"!%E +xԣ]R*a26z*t$jfOsjxt[i!g;E܌ UD%0Af? D2m(ȷަa\ɐkni~56Ȣ S e5l#!P!O\/$thP ):%E]H$ fLO BM U.T8YS8Gűr0a85++wBdަ|&31 ȬAxkuSߢ7xn+H)$[s[Ϳ}Y ~&cPTOaK*)ZVjG,rE)PC[2h(Π Q{G' aן՗}2($T"rfeoL1 .t׬JjRjeeY~SUQL;R"u𞒔:7UAgѨ. i.@K)=\.La5 QhGTܰR$>}<L߃={D dMzdl&".a›&_2liB E_*QICJRbKiܧpzm_h\啭WȕCq9O^b[7]7.>='yH;+-㼲 .U)PaJaEuF]^WdF-*(~Ӯfnʟ0 .)Jc鴘q&/1nr¿vs[sP1֝u;~>}ZW n;qr=iX'"!N:Jy*+5CAt"W'`ޓWb4du& &D"-}QH l-k8}48 ˉ90C?0en5I g85=m7aPM`S˥CWfD >6Qljzpc9X%VۣB2̀0^)b@a~q]J0M5WRug*ᄑţ.*"dDQ8DM' G whl"`Wr(أ3(Ih1&6 StbtUQ3X.9p &Յ\(h:!dČT{1v2Z\A`\_Gc;?n-٦13hAJ&V>Z1ky0`!AD̾+ǂTL3}B$ʮN4:WRB'$ f2^= x 0 .,WxL_t&aX!*u'QV*GNұ|?kT֦Q -a`BB$9ç0rDI~IM&T|=aQ<<YW%Pb6L)Ɯ ~Ĭaf4z$=NJ6gZ%qj@2i6 CݖoY4\n2R ,G|3LXm)!$ {G?{Wȭ` 719͝Ak"@l'ԪJL!RTɒؘj<<;yx>lK;lB$'oo+>HBÝ޷8k'-UPffhEȸ^wA/Hh0S"S˷k\քU?ӵLg7=O^TXv;Pё=Ahݎ[9]EΧJ( &B "x6$&ڻ!UAawRTThS@p`~1 xpѥZ+&5o =r}iWh0PTs/WEoR`zX* 6M7IU)' Dz^J 2|*fdi*SK߇C&'6@I%Йh*eiL.7ӕR݅9R6tT<`#KjtV閔l~xG1}G!q( y̞(ɇZ^_E TsL4+^^`S_WٷlPlZ&d0vɈk%ѴҋElE͋4gd}@-*VK)@M8P?&q;g #+n'f5wchΔgGQvc498_딉;&F gӞ6t؛Y2t(a֫BB1ƪg;O&JJAPȟkυYT<_;[>˯E2u+֍]Ît K%hdW;_;&^b^D8b闿vvi菱vFf!:Lj621[DLިcrΠW:㲳()f K+/ӧ 黷ߐ0S}ŽH2YHk+L`(EY\%|YjyvF^.eI)È "PfANˍ- T˭K‚YA`4 ȹwu( ܄fJ ^^˴6P scx&e%cٲns3 2JSe6&ty^䅑 ߺA _Πn6F1̞1n*ѲŢX uY) -QNntgvޡlj{ !o@SE-CMϐeo8;UۺF[Vv0u=UH|%UTX2fbVjvuKiW4PRɕ.1b( ֵ%2Ig󏕧1$ fߚ t澢;D`x5kSQwnre;L}7 A20"tN8mźKBK3$(E@*ݠKc[R~5Hؐy¹jwp¡AMEp5wv֑܂nyZաN$T8wT9{aCTmN4}ybLud,])^m-ǺYZ %(?>,X!Jq֎+8 -@h541(j)>3B8R_Mƛ{Б~Z-[ɔ ?z6ou`)׮ #X:Om<ߝ;CҹGre88PlD:,jPFXh[x#UGG3U ]>Lp3:HXlD<(`_N(ƒZ,MLQ\C\aR<,eI2RyDማTf^Ϣi.x$4pi.r? 1P PJ4nD0B ,F5ޞϑ2i}$ʳQUߪl6@`0~@>(z %4x Γ;[2ST%ÉŅ\R25Wzye@8bt k,"ZY,by.,k7UE ѰbuJW(/( }wP*U*s?&>xN݋4,vc\5AXH[^>rֻo0e R`>DH(Jļ oUDT HLlD?z(2iCf` F_k鵂.o.@I>UղڀP%ϥ2L>< Q!1 ɊuZj/^S7u4'&PZRJ($"ږPd:*xy&t3Hћ'/BL~Zew?@"5ښJkO~αDž՗ّ*U=p44P\mo>LVe<:rsm! _+oYom:7Xxl]#7U1s,Mq%zCz`4q&qu&T  IW\Qzn'&9f^hp3R5He;]jKTa!}"*΋Q5"Z,#+%&ǷP1]cO: 9߉sV?̱’p6LĤHgD+rl- ֏3 |Ǩm3Jgw@mpR;u?C͞ -WLqpwٓBc cDnp qV8w;ؒeUq.B*8lx<̈́QD#)HP-q~rM JxLE8_9\)"DⲜ"MTyKP&naIL>z64V NnmUUVfH1 hawXI%)1JpނYؐ?Χ1:PQ$ D/QH05Dl\U][s}[sFt NSv5V&Μj ]a2!m@ $!FiR1".ky}< @$, 44|<lF۰t[HԀC&Dr$EB((Ny߂a]jmeaIϧa!˒碚#)G[Qciɟ [ܱXrt< LCg4u&xڛ}$#7Ujׯ٤`KQo43}oKJΙ |iccF'jXyœMڝbW1 +Jwmm#YTTK~X wY[$+b;S,S)Ei7qSǪ{ ei#ײCʾǀQÇ86SA|ǣ06.} [j Y,;C?mT)#/Bĵ؛1p*٪,>{ũyMa\os GGק]SѵYGR\:>EߧοdV\@ 4O#u!ൗ ;mTLK@GFG?tZJ4r])4{mp.hхݭJ|&קoaMT;{/. ;mTls[bR׿ڰdrŊۉOЭ;"@Nmn8Հ!omQ&[z|#sдm$>k==ͧyV`MitJ(|i{s\Ƈuo/>8pi" 228E++o}+q݈/z%AJ%LF`;. {j 9*39M< ҽi=;,*(2JE"jbFijNAK(j9-OQnM@Y&,,DA %P ) F>C(nE>6<*ZC/5o HJIs erT!01 V@'xL<゛c}uA)y9̖!h0~ù{PƩBJz"T ev2RAY;7jt^)nsZ픁Ix tN;qZoiC+Qg :GJd4{F2,Tq;r*)tJ@ SS!kF).|!"e[JrD4."k/4rR~w{?fnbW  a*+EСj)s̸*5Ϭ ı( L/ 6|(qk)wp[ Uh8:`5>Jd!䂪K˒(f͞p S$X_(g}odh!vVX}w>i/G6 ͉ܯ<^mNV /nʹZC~1~"XQg U͎9 X;Y4$V}YiٙQ{/dP+F;2B)b*:w ;YTmÕcC1Q0b{ëhz9ƈNb %V C.wDĦ Y+1%> [6E6F1ORn%З]r=~.l& emYMfeHZiz?t*2$0SӿU&ԥͭnb[:ԤԒ L+To{'Vi{ /c^tIRBDX[~iZw}8g~β=xI$JDj2N% u{Wuq΄y.E^"/p,$ 3LF9fkV>zUh?utad\c)홫 U. PTy .߳P}9/^MybʽLLSա^"HZ@qSpw$c7ݥIHnXJ#],i=cR ㄫL 7.?F~JjN]ř-E{3UuM$/sT'ĘY.f"s u\ *\f18+Bt1vyyUafu\zƫ!g˯`ޤnH+d%CP.k6Vۓu#{J+&5 O/T=9>RyWNoY< ?c s\Z,~?řc> 6(5? :]?yzbE?i0τ Ḱqw'qPǸ>~6*~ə>671ǝ5?qO5)s-=|{?-F?G&8pqأX&7MWsL0X?~͢/@ja4x5NQ|l2w [kƌf)>{%ӶooԳTleF8Y,9׉0zWO 3$9m, 1^1eeƊ:?Y38Y͜>Ρ  UڛcI8G){?O_Ny[ +4m |S?.  DԬ%m8PFjsOFE.{P%!6#Q:&Ⱥ%n[k12<]_-R ϳiKz}L 6Bv9Pzˬ-L,7HiwX%ABcynP#uXOG܇bJFLˈȰJL3d\b5r"w0!t'9I ~Gfyk C1Z Yp{~ƣ7;۾e7CQ+Î;2Mu}QWrnƂ!'Hsh a\?u#y Nt1E `fzaF8qܹ6NJ i~nSoʠQ]~l*JMsUv<Ld(%P Ϩ!L>MB'Ì"Dxl͐u?`s*cJ!._3smlJ| .c Һx6SD<ϳRPet/T6VH E $E)0&8շVtv9Fx|{-wPrIR*^<<(JB: o_<̞%;s~tOA8'S, / _{ݯ|!ꂃjM2I/h/mu#,ӗ.KVnN]Ni^\=^X^]Z0AU+7fd5 U 8i$Di72*4ءg89>Inh.k P+.ǂĆvp$npqSp9AUlo <ΑapCZ~rDR~ڬٓ[H%݄`CK =TX"`a׷7,Z,6 M6`Da.sǖ1l᭳q227+ ;CMiC) u6 `كJ8P ' XxQ|򍎚 9R$5I!/mkIrVb/uģS1Qs  D,疣 ϳ26M^Sd% nc;ktuf/`kdkDSZTx@R .Vf鲅$Aev<}peWONPu.5_9nc;w{\Q#1uk ~g宏m (x0nI~tz¤DbdָQTUb%\{,Kx˔AK%x慆t,hw;f9quwKVW3 D-Kv<M_>̷\7e7d]dc^87h1|V7F>^2. U~: RIm,G^eS1i@% ONey0 D ܉b4z_t{&HMtXz3$M`Ok/1U%d-E%4Y~tY|0:*-u2 FrmuB>#?ux)qq_<6ڥ]$%:)ї'Wkso֞:\hyW.竟 f3)eT8%Xb XS*sΰڽ\ȓ {'(tp1wRdVaf0ʶft$!StOa9/ZJ W-m Y;gβpCNNxR.\>6_2wSJB[^r,|Uv^ȍʲGS[+^fQR#F @(#"sSW}Ddc5Dgw7!C)"M E:,̎O.fil+_ O)uNFO<ذ @ո<eTF)<(W7KTH[ ,:[ v O@6ʜoahqe:LV菓ִv(DFI'J0D 9DtxMl$TE{bZRĕ I4KN'UȨ |Ţs J*`shz8We34Apw`M%N{ZKVXK\-JBƛTC^-B2[;C<̂ߘkknG Tì:{M*$mٺe[OeJ"%PiR&эF_v77 1Veѧ>!6[q{~ >hڎ܄qr\c~*hG—'[+]0`]!InokX.4Πחq 0`v]_-OebS%fXc өtsG08SHY|ө'"^0zLj|5"ZN gl]VNE- I^\H<6l`*'^ R-µqHXs2H6|/FSBxr i՞ ]Ꮇ+j#g%"c;Us!nkT"84 5X`4v8d7#GsZ|Ca L7 ~ (L|^T.dw6]; Ag*p(^c `#E"\r͋a~;bY[YHu#7Bv\X|0%hUjW88EńÊ;-źed@G٨FKQATIO%cہlB8@&Ɛ-jWk"@'BձpU3>vY:0Y ?[[%udߍ^A.۱⟮NQ&|S ^mlVEM"%Xl;m43~V0Etw]Mch "ơ>3 #6H92rB-4lӖ;dJB;fO,5rT3Ce&B*b#%9%UKi<Q,nV GTqF R"Z'n?v3]6.uBO=PEfdG^JSF_fZw,6֦B6 n^Q;{;N񝏁IW|iкl\b4.!M.w NMCd=-I}V9VӾ. a~*y$yuŖm{Gu\~83xuZzK7WO)#?h^#qTUA 8!KMY.NCL ipıO9MZpŒFRM8^Ug8΄!:AOe4*vw"[hb:gigv\Cl6ʅSΰd~$ g&^8UN_٦x—F=KlNJ2cDFe`3cW|kUmdK2b8VJgL+hy*Bw}dVa( ꄆcFY phQI*h{ƊH\+ƌF[3UYwĿ݉oc4J+ۏעh:F"b(ʖ;h(yJpgC/aR Sst݄]ާ`/@}a +; 3Bcrυ`O;ÇzQ˕պ៽(%p4:Yڦ}֫;Wgokϟ`zӲcM@VcAnt~p(7}: Nqygg{[|]o].5W ё%{\[[cQL0'A\0 Pd7V AaL֕\0s&?ҿS/I'#om' VN&6pe&aM!Zuզ6_W InyI KI?A9矄ኀ΂qN󏱲hƋ >^d(<[U%וRJ&ڠIlmd.iǗk [c_TAs,DR]Jp)^qp|K?ƣQ* b$|OKH:T W`k凱]3m1s_(8#N^}i fS_YN?\a5/O ד}VoAZk֜,­_RGI/yVSC;Z%\UGB\2K^{ێ^ v+ɻ鞸:IK%SN+|*hWKD$x 7ƌgiȣUoB %ZGyB~cf,mӱȯQ0 Z]_#Z6Ȯ֜StFcvհ9E VcE-Sdp`u5,{i>hwdNi %GBQhdGDE 0d>b-I,% %4J'+qӗKN5F.2tpg-`N씒H>  ZCeNXٟ<.~[7';=Ou"w~^+m,]s><ٝŗ9XފJ?U*k:6"܀K1P~(Q-t9#I1h6Gg=?n~.95wЌy&`+nDj pu4]>L_2_ vܯSybsVԢ•[;I\6!:U`nY&!34 &ׯN.R3}/N yđF,)md^a*hTd)CV` &"FRuG׳s1ozMOrFm[ޢǻcm$UaNYd1c'S q9M?]9R<nR( +1sk;)#I(|OTAmh@Wfo_c3"OE߁Zn&ںhY:?zmssL ŋw^d]L4zYdd Ҝ@%q-2* =vIcz ;E&KG pKQ?:hLzf&U43Em# 'l>՗ֹb ;K"i3( Icu&]m.!6ӻVfc& .Q%N.W\R7Zz !Q1QVМ WFf,0IB5wt\7э;vш `MUr v]H6~EU$G"Rú Env歁-= dD=;}Gm*hBX !;ĩ) ǎZ hKf/bubDq/.|; AT#A|^jz`}Npk=zJNLRo ͪ!E(+j>S 9Ɛi-yɃj;`Tsǂtld]])m.(n]_D٠i%b.cC֏9%ûZ6 ~ p >)G4jlaAId ֖-GRp&4~VA&Ԉ ~k-ՋVDʮ1^T \tt.RZhzd'nNKڲրrޖ.&ޓ}xNDk !vV.A㍅{4{r?|&:ԈrAp(3}>4 E%;`,IMP=Q e;x[hO- _j4;4>ֻ0R5&W8~[pb9=mLP͵V0mfsI^\רhLh*k$u; I՚LS=p=u^jɄ,LH $(OP*,1u4V|Z1f47么B~u%f.qܥ5n f:y)J|Cnφ5R{.4`^⨬nq}첪v+ƾ82+6쩎1ISbF BP "A4qB:cp``١Pk~ZU17'&  dlYB(x! 6} W7`%dMB{(iS OץQ+6-(i[m˖zw.\̚i1hl%K7' Q,:{sO.*YJ'g4X~$gs䠵 CE0&_[FR㹨4zj'p^@L3xE 21T@I"@'$ڥL5]5`$C2da -Qd qo\O,N÷c?_~MSf 9 t8+v]F6i~?{ >Fˣ725$15:k Ȩ8 X#2E$S,0.g8,]u=$VdٻFr$W4mއHa&=0֪Q쩀Y#5=fJQI8f{`է z>rle bـ  n=u!AkYD'4C8D{ǎcO\2E Ց\4_p}A5 yvnG׃_gv8A {sOyste8 }kD,KmE򭃇eXF|uHI( iw`@QLD ǬVXL BJ!QFϣ򭳜AKqbK2]b…p^ۦd@>Gx]̌Ãgo s^KXR0sgGHMQYhAAGI됺}"5ƙ5_F0lMۅY&<4_%z|B@f Q&㙿;dO/J{ȭR*'F6Oi)HQc-C0E5Q̴>c]U~WǬbU),xB 5Fw<GKҁV \ ))tRӗr'*w _(B`-e%*iPQ- 6e!ư\Vz[:z)1l( rDkDžhmZMDw;^NWf|Y"u la Y=bkuFZ [Ub+租|8CaCShe rVXb0~DpSg6,ԣi[;e0ZT\.Ua 5n$K;Q~Cr] N\,9*Iɇ_6\ՂB7"'?@?#X]nj,LT}6 In(V2>ˈ#Xf ;"0 x*gP#W{^ȍ~}w0ERt@U*6&FS}3 Ư p,۸Xw!(6S$CSfᤃt^pE=fb({al8ప9r)26v ^.fNANy{1^T*Cxbg*J9!3`9x)PTQ*(v)z9IU7QX~cb=t* *p{̘iXFKw=X6A/4~N} T+F%Qwq2E>D8phӎ|l %u# KkdA! #S[]ʒK2 AEDOAftQ-(BW_,X!DiØnH&y7DʖJohSMk&Gߴ@)R\ebJ`"{Lh?Ҋ̵"'P U7=&h8OrӢJkK ADZl|_h:Üo'svg F2JШii瓆h8 AngVYt#n-W8Ԩ歉* -eDZ|PPDk,W##x[c95P썰Q0}7b mUT%F)])@ʎSa*D (PH達Jkw +xO@Q0֤JiF)}!HyNi{oQX;S-e+«qT|ICXZJCfpXaa)J10p1*L_JDXB(QrtFNuuhj,Xe0/dP!ӢKrG02yMVEG%d!,D^P]%rzͼvUbc y7/+YVY*̲}oP;`U(CJ@~` @)Mj~X*Nkp.uZVգ]s;ޒsSJ|5 q4ˎ[3Xչ^ M޻Mxz0N_)Va0zg^giFNߛMt)(܃3-"ӎW P;nz#VuNx`%'k(& #֚WFd0JHa[TDe!7~jRse<rr:[9ݪP;**Գ"#c"n駵ڌ ,ZҔ4'[s}^_M•t/n%~9=Ti7CRK^F iU 5']֜V4S*aJ ƛ֪)g C)V"Eo,[!$ X?AKjK ?y@ӡ?o&FvO4I*ê?zUl0g0٢˵Lu1ZQ^oK/qFװ>ez?؏U`}?O݌:l#yVRlE8Iamqtʡjb3hD1( evW{|ι멺R2&66έOwMn SvNQ2BƸ2pvZ*y˄ʗ^=qW.aDp7NDa:;J,d9jkv cp_|M_-ۣt|G WgIDT(:]ɦN Ƚu^ƖX7 hEaAhg6AupJMhr7nlN6{/92Y%Õosg/k[&3XygY*Xu[z!|tMFls-;Uwչ;O20شЌ>'U{*Ք1I)}Jb? nN^}K l=-./ᧅY><$&+$k5V* =˜3B B3iFbd,k o"" JR`xxx|o$3c_T*ĊЙH r;͐ўD"4{ҙNCҩH̭O +ԯgGn8.tNН܆xd]*^JjֹiJ47{5.w>_xFӎWRJskeyuҁ3T8T!$V(P;XI4]@Gm }.. 7^R}/v]| S:0jD&Al:[fc/`iDdQb^ą'Fj\Xǣ/H`p 8Znѐ5SL*U.UeTO:/뉺fgvP*M.<H44nngt5@13e{!!T9-*.B# a. ǎ1%7;XΫp.|S*eJ`b˭w9+>}uBE߿n.t=VIC (O`sJeU{ki)M}5MuuK\9\۷]<]zMaFF?y9s6Igb6#` AZd'Brm(l/94% LU &Sќ/l#zUX_K\NЛ,)bL)tX/6bI0@ (ƜFbS 3yfqq;= XT=}ZY4YTIrP-URMb޳#W}2 {~ lvD-a,֖FSkq& .aAsټ\\Hgc4kMݖۋԮ4q{㄂vkjXhk)7a`>+EYL3E[f{I_f1CBRhnhms*VR{Q ^չmm3 ٠OR2}xeJKȂ4OTLʷ9 !xmoܠI!2rZV/BPӬN$b#w#-jڟ17Ͳ`L㊌= !,Ks>J?ICCc#k4-LYl2;o'Ϙ8aTJ]بVJ%%*nFtԢ@R=QZ綠1LG0w*St[ǣ3tI7JYYLPig(ϕVƢpi @%6g.ױf9:ë@UmW>5sX\3rL N_uYMs,/È^{~F߁ .ƒsmefj/}Յoz,Z5Pg~c՗^ݥ9-$ʜ&+1@09_i>(Mk}m!bh9ZT$mޕ4e ͽq՟%Bʕ!i* ƓV[b| eދ5&wl02P_c&P#;ױGtlCzQ` S>_Hՠ#L{:.돩)mgqR_^pܘ7j{6EwEu9OH!>Y !(WhYfk)-\NћM&*+;MsDVuʵVk'a -/kck f,iuGGF`oCO˳.l(:{vLxz?o{$GJϾ\-~42 r}jE\%vJ䤬В1Q&TR@aº-N*6XWi%rRCDnܡ46X1A  hg(]w`Em RMH&ɇuzzB9WI—7#>zZX.#AP4dBU (QFA)RF>-)ɍeQ ݤNӮD_MڣX/[{?g K̈́Ew.%j2@)iU2Phݫ5SKz+S{Wg1ɝf"HCi+!P01wunrn1<2<5ƭ޷ct/qxTh?nñ]8 *n#Pd<}?w|Y&vzH>pU|O6wazKڞdcjv\9͏W=oOu' (N@N2Ơ )_kIІ֔kBaQ^Jk d@˗x'>C[XPثW!b6LVѢnx5|:C'6:g 7{I)c0\7z͐m6߇7w(% Fe، h0ŏoP2n4_}oo2:"0t` f}h Q;ĂV6hʄ}a0<^" .fAul>ZFgTrujǾ5dZC>{"m})BSMpO|0[a꯷m2Qܾ1طi \ WI)|ˤi1XZˆ/7oV3S+wt[McS V+F*1^sRP)\ȀzKqkOgVoY-[|^S) J1}3|A&FҊ.%\lU+s bR=so +2hp{!6&J k ՎHU*o|,C?x}Nɻ4Ov{ YP ]'RCxJȪJ1y>LNO?Z_7,.K]^fD2|!( "e[hPf%R7 l l2paa ǎ vkr1ͷ뎸&v>?Rtx l4FQ*.-nJњa%+wxp7N3h=*Y1/K89ѧgܲ?-0LcģRM+?oI8Gfh}#+p9E)mq,wa;?&\䁒/_g_Jy z"tc*TQh2n#7 foMS>JTBݸv`fLgǪtOLZWt 8-?` ]e k4"W %O_G]r8QN`O,"cSNAdr4_@;gʗ۰ bS{bhLn45Kum#h&e(P.5lH<@E&;v +"Zj*~ $^(yXHc\:gv $~DH}rmmFXvv!*+m V.y9%9OBUun7q0sF$z楋a7bm.t8oQ6 ~Ogdoܒ7w` iI[jy_H<lmBS֑GL'߫C.>٭0@1/75Tf̙YWy'e7L~SŔw1e]5eq?MIƋ,gACeF*)t,/52 3m!eͼՃݼ-5ǘc1*(knQCM͐@!ϨNA)|R=~g~ɜ6=2i4%bEqkJBB#.(.PQ8nxI*Uvضhk<6_)Ͽ]hy7 (Wz/? _ “ \N q*_d *FK?R]h-|8E e+Y+m3T 9ZdE >S~f[R"Ѿ&ɐeP'kÕ7UR "3ym6et/'߾clWh+suzÕ<3Wj[(xI : ]YhG쀱1y2kR siڄ`"; q%C+ %~ٖB^snu#v2U8Ty^O0$LA{WhU-od\#\a8(B#!bӠyYx^+d_Vzrk!&Mǚtot'sIT vk%!cF&K5/nCE<@Nkn hd+q);<$s)ۼԈxd*xEQ|X-] jvPmZ0O}MYVUD_B+' MI简<Oa)/Ri̋:)꯲6@l$2T泻CA+Y=˴% i5W_M{j‘>WE)bw«ga7*Yk˘T39qrB)RY$Ǝ86ωoG(3A}{f,s 2%TN 5˫`8dF]T67kd} P={7"䝕l<>,.Z&,TVP`!Ew8Uk[Y1B8M<ЌNtlnhP@XQhnXFώQc]YzZr]KA)) Gx+@1m]tLmq9I5_ڧ;P641ϘTMi/A$xSdiGCUmu!xEm[֡*?p뾾T;ih <:I8<6jMtO 6 KiXYȠ9,.KW tt m%67 q{SՕ64^JIxcxc+~IJq5SavwsU^uqf~W"_CPs~G̕}gVI(ԱJ<6kMRd6"ѥ# Y\C<&a&{v)EAcV=o*(UHbp= ]DgdD[[?lP[4~n3=Fяu'mu*)RCriܟmM'x|O' ('{WDuZrXݚEyww+EJq*b\+iRl ܴ +—VԆVM v]!좛uBY.UWoߣV#uGl0ou'XEPιpOLU@*Ԩ Zyh׶b TI=hqn_2N295Hle^olA>s`Gv@s{77wR~{ߕct"1-/+hҐpF#. v|q ]N:3Zt=t LgO:S|E#jU[,ZG;NDTGRXv&Åb|--9ѹ wFՌ=d-pgؑt#v`-*wU@ù2wŶ77Zݎ\KP䔳|5ȿCA,> JEK꒩ %xB[ +>|+ yme<Ճڐ°AhrVµ,X[ukrv{7םhyS; {.7t==_H =#ά\Uj$XGs2 Clf$ 1$1v N ~~m@o=f2z+\PKvɈ7 1j/Yu8HH۬Ck#ƞds57:F񗺇d_|}ͫHU(zkξff$f_53eUV10 K`Ը<9{y#,g>}jkCpyE1^Nޯ?+TDr rٱ]in~ IsџTGz١\$ Vݧ*B)N. aȍ#^>ZBDjvna]@KV)о2@D$=?g \ZxNFH剬sf6SkҚ$f$Q +EF!p=Qfl$ f<*QDBp#y&U١p\[hBnV"NP_5}zr64\J*,J&J)XIr²-4lzhv{)zV}Qz 1qp-:`5NTzU_/oh4WwW$'i&>Eb#q4mӏO"}Ah^fetIq1?ՐO:N:Ɂ3V/ GG0SY宼g괋 Ot8t&r{6`6Iwsޱl22ζjs,қ6':z+lgkoJ)l3OO:O h>pPOUWC#ؒơ׆;j Z'`hE!:FvWUJ:Yk:Gm<Җc@K2;4'DxJ<ֳ_:_C㈐cKKS F6A2Sg fI9;ny)LR9!mq+'lh vM"hRflwRs4"+}Kp#K(i {=PP1ܦےRGS-E3wJI ƳoiV.n}1àzpѹqzZ-z 5b1bù{P#ڭ c68Κ_06:p$Ԝ=s[uHs=QuKVd[f'v[]Svh#2vjiZi1"(2v. m3vvYP@C͝ݨ&g))Q zdPv:uLpejE'n 0rE mn5uPpc;_jXuB5u'v8̸]*EC̫MURF3n(߄\qѽ+iV*ly@ΰ}ŕ*;/n>]o c殎vuRo&3#b`gɗ_颚'N[IqopLeԦ"I3~=vBwYn~kIFZ!]M+-_gs ^4:(-YL++YY]T!cPT.ҟ7|iϫ;u(ءj@ʞ]c*h-@=fB d(;C֡l{ 2O&CƲ3Kmit Z ݟ?dzDC`'Zow׋$G٣%}g< `wyli$V>K^t?#} ~ZzT64).CI]Ed"TBqI%`{-AFYxkyadrO̯֬Pj_XʠQJ[DUChi"[p,-ML'm&'K7g3;Q_?G 뗶 L[s~ω?w;X}2)FXB^˴"WtN hR8F|#UWW8;KsG@0O@GUx@ }Ć`, Q*/$DS(ʃtZ:Pѻ{ee᱒E:߯kgg deoY2s,,Ō52zMZMU]^ZHBnfg!Bvމ䶌 Vޢၗ6E&첖Wךr4=CwAq-QUt!F,hciMkR8FhmvM/SkCy ĺsцf %;)Aaqa%{Ai]D!fQ@k,~( 6[),V6c(t *4 t⽳VkKV: %’LG)FƑ4#ǀVۈk/xY I#+kǒ_;~iTU4R"HdQƐ Bc!%+—ɴpML v]wH!6U=&ZA:mU<:#MβfaXa\0&1ه@z lvYgF6%R1R}:9!5\(*_Y-e{?o,UbH;)g!`3,0NnU<};|i(ZZ{8 50u-6`sͮiE wu|(O}s܎ʴWbd`g;C4sRLEEp6~i3]6TTnCz{QMZG?JFp_M9Hl 6ج:- 5CmSml囩G`-l,XXl&dq[IݢO!?eUVr,v[τxq}E&b{fӷz6[.ڎ2N@9ɀX{M]c+-o(fӣzryyi~hGOА7_yN{M! *Ṽ`ު6Bjafk9"BLp>'zJ]n./͈y1N 4!|t.ZQ3%e1& !%SWփ{6n鸟][ `?h;ۓW,(ْ3f0yYUzC'`-BRUٯ)ͭKwB)S.]z˴~iȤҾUʝW0;΋~.Ort=:1QW4%:4 ˖-YYUzP2FUU s1b@x-XZm^21C`dg;S]"6yjTB^R[@BRw?RJC@h;t  % 0olo*ڹ&ak5 /! y V$9.Y445 WgT.ЪRx&# Ӏd\KZC_4s><B`(CY5jw.Dm9J;3X/Oy\ Ꙩᅣqb;h&4f4T&DG{2K7:|N^VBvr͆-q69z:4C7􏣂'Hd2i}tz3yRLdWSbGxG=hNnIs'7mF.6\iQwSC9N G-!+4ɶD`tQY%3O?PVbKX-%ѥ;Kww.0sm9qkԡᄒR%80oWU|dyiHct׻pdȍ>~Xq}ڦ?}wr+}(2ES$yde 5(I7 #J| HyE9f3GQ>Ɛwe|@3~b }Q[ur{{By8YcD@)Ko0HL6`^#k~UmAkͭǑe|KbV ʓE#6:01g"ԩ,/o{7VӁLj|~p uY漳e`w!<yLhUAinbxFs[y;WRYrz 4 LC抧g~op ;Uv࠴JK2`b^ 4y \4$g-kԊgY%Fh<֜`mViD/ 8,qiOӞ板 W`=yfz=Kj ]suN{vO{4~v ݔϱe,tmڬ|veC&Iٰ(1"XZU|7cHʼPMs J=T2CJbMɥ0k˽PVTg/#RFCSqT{hպF)%k .(^S`<|}EYXu :rkw5)#v+iI O2|cxG kו j} ^nO3K(xA%yD EI,9/}$8$h ݍ]HBߚ5ۡx zklo6tA/;ӏl+n(ైF5d( 嗈Dj]78Q؇(\9Md㞭fZ/@;?!p;'7E`C|kd`Fo9V^2e|D'*X"{-)L:fv/uuGI9'%nj(̗8m>\]s[)+Em K7:;I,2l<ҧC3Y=![R)l Ag 1;g޷xD?^rkBi Hn}wfڵo>b_ub^f_g8CHړtUdȄktmYcGj5 []zVEGI۬YOc:.@g*_^U*""[ME/f-.,zXc`@ct]!VdBַz/ݮgeBh{j Vi<1YᣕJ*'dT7E_W[*pǮ¥熴X 9j?sd3Xn<2|}rlN/Ц>zyRy&,钂7s~.[۷xD/A!с(e,ݝܺ\]/U?84Bo5٫cLH!g-宼(] r)QEE1mF/ *_IT=NĴK:HᙌL:q,r.k=xXcx}w5.bftiJk, 6 =7g*v5zx) tI(#Yt"iY2Ե5 5eր<HWߵ6/{[n.;7LY1,q1jG ubjiJ<q!ߪyv0"P7u3l :ŇF.@׾@j]7z`lGl-źi#KA&'(:,!qTm#t"MI/:$͞rQĠUEHIۣp`LR  tbuuvzK}ArH#)d>AfT[\\[>7w|UR9;91 +F 3gcPJCB)&d( z8  J)G"Nr|48L+dIZC߽*DfTEl=YK{2Nﺐ*IMFfMm1]d r/{It2FopZ9ַj^2ӣÞn^|BޕƑ$"bY,blSER-44dꮮH*@SY%Eڏ8(1JzL'@42qEC $"8iE){%Q-R錹:W-?EKlGʵ: g1b+ j!(JlIN ~QVlgPr8DBahcwXPyB^%ojkq8 漹_7[(\x#/S#ݝѵΣW}ũ;&b_/KK}~WC8:];>VV|;`GuYλ҂j ^պc\Ex?#yռ{ _vx9 t7t8gۘ0,:?!qI?Iߚ0dv{rK g'6"n/QO='DiyjJf)k L,>}uz`{TwMTGyY3['oMKTԚoy^()DSUwQ/ 6zCbɹ[dURPUhNcɲ^|èGZ# ٬c͝xiӜ:/8U)%$2T' ;3&7سD4Blɉ;[r: ,cvIdm,ނ[;n;LKRi5^j{WglFmtQǭ|FeNJwg v8cpdI\+d ]|L@;_|"^+ o^U,0b,>d$oֆEDUY(XcAN"ƃ8Yr#fذqqNC9Of ޢN:.EL`$ι5{6#3ۺD9D.@¯_p@Cҝ:΃ɚ+i\&Lxpʘ2f7IЖ_2vivV̱I׽[aҌOR5\rK`T&w+ [|5A\470oXȻsC6[x&u<6̤d+J-uV<ԁPhY|*:F~m;rp*?voߙDCLK[͡ݡŒq/yox΀ug5J yʼ2fx9)%Et:\4D;Kc*;"3vǤS`o.W7#ۢMhshcKe)iNlEٶqNj//.xc)>;r7XsdQ2J4v^]QsNDG'ɣav8sgqvqѐ=чstߎ-MQ |E&{EX̺! -dcB6yRr@rQ.MGp{ eQ~۹o֚V u>-l 4'sP2/՟u Z[ &sy2}F: iWKpqSbb2;| 3snڃ6ZRdbN&fɪB1*+0vk#LuBω>g9襑@(9ho/_߽)7Lm#ѻcZ9oՈ !>ѿ#3?Twg*h1T5|%NX%_0||uJTQ /5Δ Pvѯ;v7xY@J8=;9YҺQ8HKTѢb!`2!AzѶX\QQTY_@[D *`vQ**٠K"R9`JLhHZpTk]ebʊ bTE 8%5يhk3kN5(dYP*B5>)&S rhd!2n͐NS&dkNX^=AQ8_\bQa3$D9G/9w1ga|}URa#2<,曐e/FdL=$G(ؔ6[ |g催&%Sa˚,U)!GY| )`;/Jb3(ND @WFFP}[770Sb4WA*zYB uI0v@`̓@jw <ìFz 甡8L rhϿ o[u¡ T4c(t0h (M43QAALaRi${ Vl)g#9d7f`uaCǘD2(rb7#6DwmmKzY.KUw?%9gHÞEtDɷ߷/6ɡٚĆ%p꺜L2f:=ngn>Υ3;VZZ@ Q;ٛ7Sjzz"QOnzQ; jfTCNyiq=g a}}vZި{Ykح'lH|btdTaErpNw5nm}VvZH8h&gɑ-|kt%E_GWxzѲWdo7zF-*Fr߿zu?9?Nw'G'o8"OOCz@^]]̿z$ Se?4LŔuZf.}nu1Nl~~|uۭSMj;'Pa7{|@)ywiZ3wΈ,cIr}hl%OFy43آ"_L',]~'eZLb>Q$FJ IH9'r>$XiL> P2J3B+؇,G6 <>)GuOOtހ6ݮ4P[-bXF3ϼGa{^Ӑ-z¼f8$'ձVeeWZQ168¾ ,cxmTy1? @-EE#LM ߏI [sY{3zw6 nHV-z!riK>[C%ڵDbXgs;77gHsn ZCl,j"drG8>n8qm`*\ST6;J"X#4iDiYPI!Rw+zBJTEjHŀhu|2R}.U 5z+}Q{s7Kn#mzdgJ G֙efl׽HeH<_}l=bbs]VVdj1J!ċ;Nˋn\*F"M7ĩy)qsw%5K:t\}fޕ0sQ̧<-#Id,cBf,T % -BPY3X3Ȇ/5Y-IS.F,Y UXe3iqЩ<6hgT! W`%|m\k胞eh߹HES*}LZp1-179n u~H>xLuAg/N@bk_?>3p;çy vpAXN߿g2^ޓY[9VsTwyp7[~8B9ѯ^I3d-0?7Lìi/k{r¼K}!'! | "P,Pܪm:Ġ% )7-Gb9%y5Rb*,KkBMI啶[;Q>HS1g>M껞SlQ}7}&M6 mAA wX([lvѢ/:CuUJvYe\^9~ VƷ.EAr֢yo&u=E%B Eh"L&$h@[5) ۂкAHAB*Rd,֒mޅ˷bҬϝ=)6vaf 36?^ѣ ari)6Fe}8j7IlB`VgEB 5ǭ3jJBQB+;hl ČĬYYJ><u5wM]z琼b.[w#drX$ jM.H<,Q*] 6[}\˗JR*/ cNZsc09JG7atuC\z=m>6Col_t=zٞi_CRJ} nL\"(C\E]_HjMp1+9 @aCXZj8OՔ~F%nCO:XM0P 'T_} <-ؘm2TxmX4<T|Q:>=_P[ ;W/o2bulG\6i֨mЇhs>3J^Cy%bfa"r8\/ZO'NgpHtYu2) !D{P^W/ m9U5 kTn,vwwSj\SD]St\kθ  HQN!"EdIwC"6h3./bBsbŨXɊ ߛt Z y35X0PL\H$S `cNPm g:,i@ƠL;6KH6ǀV"I?K0vp.t 2I*$ 0"&,.2\w~+[LYSI@d1hΒAB1E@:5%#k8d_W eArޫCFM5O$#7k;c+Ga#SsqWa)iiXQ ǘڲY˒.$>Եj*Q]AaF~MВUd \  9Nң]KsI+\ׂ?1[Nak6s[C^` 0j$Rsrx Bʕ+ޕE !`Q,%~ePWT| (^U!Ct_*]?~ȌCf2n?tP4daÞ%2 NF;2󷱅9o;PsbnC_*?Mcw Z|Hhyb]3O2afÜY2f'Rmcd=MGI|bee}K;zܣI(> [<C* whXK2Fi"%k~;KmE,9pJ0( HS<}CBq⧫O&շ/.vZ$rJ̇nZdq` V|Fc jB8(IpmU B+&P5/HTX׋70Sٗ8zxBub uY }QfZ9Ljw$0\]P[7#6dy"[҂RtrGoao8qMJ|pwgoe5aг+VlY-_1b#$eJK>UgKl墆-Q >=%ȗ-rUZsKNH;}<)X8*xl=6Rʙ}/ɛAq7}n`?:ڇ<\/~Tw看LN P򠪊] ~ !+1DH>QnR0)'E&o#b}Zq7YSYi!f.-9xf~9Bl*ſ/ʇ3}$ fAa)qߐ톼l4@@1Q~9 MZ%M7# S,y7g)ш8w+Ɖkz eAOqu>z,;?#FXZ'Vz/,)K[a"+;5Uq1#w)~4\S,3ukΕmJTĩUS;׍I\G$} 4NrSMFG3 n x2M|$1')X-Hqh yI= Ja!qamް@G%0ZFv`@&;E/ZN9Rj> ex7S& Opq ? t +֮+=E?uO9;EnǢ8J)MXlP>b t<܈q/?~3ϓ˚!{K،x2d{'>HÀQ b;Vn8qM?{ F׽G8wܡ9"녈Uz(BΕG*pFHq2Z]PFk*[UhJ22طo,wby8ʨT67#{T8D6Furܿjl4g$ѱ<:Jވ AL})ktl͛Hl7 s5g͈]#uVSHy+ljkwֿo}c+| J {R0JZ+pe =|A_C>M<8}a%wX>Q)W 'ҷj]v}ų_; ?#EWף.Cw˴g缚Q}G_Ӆ.qW.;ta.}^ƢX|>SQ1Wٻ7ug,[W .%_X57sb]]KԞS ! B͈''SGܕx45VG ?|Ƿ-c>yQ:||=|{:>~eb9mWByGjv~NegߟJUzӞ*wo9st9ѶݳW,uHtR>kj4Fs_Rtކ~n]txV (ͿNWإ,p4/5Qk r9;0Vٶ']Wqb9 &FVIMǘ3Ԉu;B1^6 5ꅍ7i3꠯G C>)@4k-,u;fH)m֧PN)cT!搠Qanf?DowdrYmj/w +__okba."UԧܿKf*k`l:/NKEEU2pkѧmnKd ]ml'ZMM{EN%*|-#{-aNbÖR "ګC*$3~c+ RITYuoKĪj)q?!aЂz. &O?|OW򎖟7#LXB2ah(^% 5 Y NG:. н?x2{V.,zВVljZC򾧉$Pbw=y?-gyZ5s"lA #e <|6¨cGfē0ƨZ'&`9ޅKq^Uٕn4FUc`ېh tMar9Hn16Il+bC9ĢR e&-4qUBHhK+_d~ G]m4v#j X3 }~*nuTr6 /?3u"͈'gYqƹC$/[9SW@<0 ")c! [E;L)8 UqbRuG [B3BuB&[Mc"H9\#db5T@)V 7*+Cv F(4>UR-WgL؆S X]65 zb "ȗvz-˅V)dOfx̧vGc]^O(1LHw$$I:!ɵQRV=bv-kɧRt?U&k9X+ȬƠ:Jة.tk!:Iw9qidG|to}]Ftl ոwhfA{q4d͊HRKʯ!A+'l%*(>)f^),th͒XÇ0DF-5o7og#Se-UcGFW|70 *VP`ٹ@ #)ڸMNwrJ2ZQRlKJőS=Z9D$CED\mUUG'j1R*cL^H bZ ȩl]ḴCnh+qljd{mg$yymwg$]Z2v=dK&[շ-6+;6hA*4>>֜_5Ԙг3\q-x2C|κZp=Ke Y jaZȶTC%?߱k*(`h?k)ޓe]l7%:N_iD>7(s\ iJpðm2sd#4ITỺ"U. RQS!T]kʱ)rrT}/~"]+9nwM/ȸUȷB}ba"r0i0^~|#LDNz>wxx !)&h~={'[#ۘ("y(zUG[9N\SơR/YZ!>`Z.tƩ b{è>eUͿ.c?{WǍo6,!}}XAM˒Wc;WKK#LkVw,>U| 6*> FD@"o`cZXmC}9R0Zu1eb؛1%Pcy4}rAvVNo*.{"UtRQ^;r+CʅD% ca 쁌#b^C,SlQF1u2Z P:mpT&|1W?o& IFeAF7 ztD4`,Vb L+ۮ؀2Gz341}fy<]s JpˈLBGTL m-VփN{l1* 6NNI34ŬFw$>5 "dxN?g L/JKt c Twsn|}u$rVxxzዛY"1a;7 0}K<;j6 ӆ*\h2t9ecɲl`&`:]c:%*6y31doUV@XJA 4V'C~1Q2ފ[?A8v{ cC\B m1Ĥ̩sVB»OP3wTHdvk G+I6Ǫ|hW-ox>a Do ҅F%];lq!ŁGZPd ARaB}M:.n.)lN᧛g+ Mųrmb2vߪǙJs#$/wQz:Qcj<=3yv.gkŬIߘ rt܂Ϊ%ǡJQtYA廮0KE䩵nO%+d$l{5,\13YG0Wh֜~;gM{xxAt_?\Cq)%RQi`,FuYQf{ؔW.1ެ#x2k"I89ި QYJAu+wfAq^i>#ΨdxolLJ{(sha~Aɚn{FHl:c_CkP,hB Kuxa0<Wͼ5EpM*d?*ej6|ڧ5̛V@Ui >чuroc*}6g"~؂c D2wydz iSUʼnyoA;vCiS2.oڄjJ6$Rw<)E`3)eSRӰ3E(^iuARa6˫ʗVh-RuA :hެ1bSpeш;M#w|9P~y%Z$^!3oo\Gb73IgŌGGNÉJ7$e 3-5"&$A\+?p-ӫη2/#VV_gJҾjӧ­´͋k5_.nx6` 2?tjRˏZ 5Ы~1>h&tCn14auPc'> v{;CߡVRD/J299\r캔χgED/S?>P/dq`Tz}5;η9 oqw\"o~ euMYZTPTyt؍<}{ڜ~;dL~*ceE,f~^@vk&̓Ojt~̹VY&K:q)B怜Kʌ9{6]*.&Y»{W]l{!9Ak8LO``ٝ0=Zl=hy܃ ךփ Tk\OZIː}C̙Rb񾰺Vnw<-dGR0F!i\?JCuJ-Us>F%dWċ[k;MĨL8< JS̰s1<ͳY& 3#'~Cj 7efJ]D-M?Tt%0sZ;l搱t"S7Vx0G̷[sjuϨ;P!<-V6JCtbW; ${>92|S , i%zup7gl+#8ȟp,W({鳢ľ>^AǟOlOb9?ξ]ӆ?D kg?˲f%)+TpdAl [QVZhhnڢ .g: bTHR  `iXb:׭Xztsߣ&]Ӄ&+*P6Q|b1ke,he2\s?'"}~jW/ iw\Αc~tN^V%Oo;c۔ qnǓw7Z8LE8)m}O RıˣWRME]"/*eJրmCH /\Z%E!QT\LɁ mA*'xLԒS)XŞ)%.3N"+$7o .ex "悜X}!RS4CߴsqX Z[1:"5mgM"7 Zl=0)x.%mrى~m;5E񃴑I1yV0VQ8Fz#ĉ"0޻qjsDwƽA#q-1[쑸Qɻy:8N6 DK8߈M)li)|M7Lmu[3|(j{ZML71]CڸxQ%u2~6F@l+@IWp{wQu;DًͲO]O-#~9fAi `*rKn7}K}d^>C4D߂Wp&L-e*_K'l_jI0B?2f1#;KDO2clEmv] c-d?w6ȭ{"6\{%F"eu\ancFw-OD5jٱXǘ]h"ܧxd5A l5YÙX8+Rn;j7ёyo w`#:Ѐ[JS2Rzob)O!jJ=:kFo0P}e-8*@2br~.}% UVpg `Qg#ﴥmX#u_VCdz$*`jˊT4j5x5%>7@Y(vTnvd+d(h(nk>گ+aO}㿽>Hh.ZBWsqw#ȗyEA|U̍'&ÐG (.jr卸rEƀ%a:u rN`]tk,3M޽!z@ۚŹ1(2qK^ɳl5j&צI_gg?5Nm,^ը{<Aצ(.O7,M0#VY#}%j;ms=9oCR`dٻoJh(7̇]0,ˎmwhW3n`F'RA~ 3KZW_Ⱦ U0OO6fd&-<0kr\䕦O6LpndPtWPY.=asfYNH7OWMvJI}ךm+ [Tl]ߍt5+H]y𥘅H?ߖVs>7e/O^ OD0?{6, /NKu H<&1:-cHʎnl^b٢TQj#fk(ґc/p 1$2XEx")Lx&ц)JapPZRܸ&h^$8Cϊ/j XlK-~>͙ft_/pn`;'s((y> z;xojRGwNΉݢOPJQ~G^n9K*Xƛ\GN^@]$0HKafr:sXo:Z}I=Ix"86I$ gZE?LTH8" arHiy11Xލӻq7JFWW߯Pʜ+pKssZ%)+NV]IV{lD-%CzS]cv%4XQP>̞>LO(ݳjh^}1cÅ ж*5{/yy6ɨK]wT*DR VtK"Lb7seZ\٫0hZ_UnpxA]_|#P=' RtIéb}rI8IhiFHsXA$&RVNmsn}]`erL)Pke'Wx+GV RpoCMשouGm3F8X%^xoF{{GNW~C`*1]8{C $nZ%YP%chCa[GG Ig7߿+?=7</N/X, &c8kD/%|Ŝ,e!\ǞKYc!Ez22GTQ#) {$GV}Q$en[c|a4R% D4# D ʮ%8ޯITR?QG:JHG~b%PQ+ "˱Xa$ VMjTt3HtP7x P;VH-66xoZy1GjԒyeZ̓v[dG)ۢo5>#%KjR=刀i1BƐ֙LgzPZlQrm\j;4GG*znjbzE\"AJ]{1㗉X B V5HV@Zn*GI"U .ߴ*$?xiwOSsWy?nRkPN1,hG){rOuN[bC30bUw**'OwNh+:E T'`J*r6`R0 hUVd$ +#T6 qNYޘPjGS*?OCX5%/'V b=o*W{]y ?rq0Q@ Q -4t7v@ZF 仛R ?j<ӵCEj!ƘށYށY{-x0[8HK\UXa06H\byŴ* TgO D8nTkzF*niMpHЭigr>&-tϮYQQ^zi7ZrQGr% b ~N>W0&vh]T^~rJ $y\槗?ˑ9/71,:m3R9c,LQ2,X[3@]ՁhX; -16 A-8 žGJ14mzMZ5 *nY' r"1PpȰ{fAZYGDaH[!8jBbY-|]j>fxWUnnG B{߬D,y\EHҖ7]`Ua60ŻK`LJe Jo DZlی8B6$àKɧ2]=eIc<7zt*r]" NnCͦ_svKTR,eR*rFB<&&+b)Tm;w'sU,&'O4,bwf#P-?L6T2$"mD6΃ 0p E*F5@*g V1P"\P:@NDXc;Ҙʽ+,5r|`7"u1} 7pI5EE#Q\wutˤx$кNI\+JMZQe U_^՗WmxLrGg.0wByy%Id앺jWScck9)\>_ixu|";*b1[|gԤAwM;h G ^$; ꇽ ̇hndzfR7ދc}d~7i%swnoyrE'0^ ͫsQG#gI ="^_.,ȟ4%)za\a,L+p6J~_g @4 P+*~X5ű^ ۴ cR\GjPY4s^eQiLE*|C?),b8ga]y8,zTK EQq iYm$lyş)esgfr:Hit{8IQ;;SPD\ޤf1@Pj_x,ó/ aP/Aݦ|J tf|!yYR ͯ8gW9@L\_$EqjWCAas6'5B0} nׂ M)-Ilkʕca?R*KSNڮo~^~W_~^16?)d~}rr%;C0'zN>82)J)t2EeNYne\x̰ sAFItX+zn|m8oG;7]-e DĜG  48={Dm6s (#^"Vs )3N%PGY!X&cVHbEʛ-`>U޼>_<>^Rg]6RcSiZfLxYFlas,g;ָǥ˨򩪰C\M%7KG#Vu2U>O+oΔ.u<ڇJZ9;Z'PP4u trX!SJrqڪΖ7o%L}kKaØ;]*67?m OROmWyW{I4 MlSuj%lXQ޲Ij^.VB{RlY^ܽ*DzE]̶|1VqRwTػSjRFvTN*Da~W҂kBSVH`@K݌vC?8O{n\i;~X+}d$,3_{r}"qMHwtZ<=KwO-䢣ߵ.=9VVxF *.۷o E*-cEެE"ON/Mk!ji .~"6G*9,Jts;8(rlar75*⁝ &H'TT! )ઽfo$9HWAV2Xg/7vZ˜q4A>SM 4ֽLcYrt4@"F:2w(sY.UHnumL$1X跃.λ >ySLK W"t j#8vv!wQ`<`7Ψ[&2e`[ *i#^l逝 $ xƍtku3hfaPM{CnX | ʣ|5_5k.|RCsyU U1`QR*VيUZM w 5w`@Ǒ`;-H*e{Nt&X5|lz(3F*\hT uF+6Ff3.R)p,zmT8b5KV!9B,OAcJ4i/ Ұ&XT`Vb1g21a 8&2O˓!EqꮓTJ8;Yh$GlFmF5[a#Jn/3Yy\e!+;0с/ع Q!H r0郳F:R1N"),7p|s+BZ FzU PEa.<;Kw ֘bH)>J꽊;bd hFE+۱/f$WR1~=]$7LE5Sxu_*4td s9|ўpk~U;tp8umܳ:rIb,EKY< FIG̃ 1؃Y&^=8k_\5 Z(U"˾2'H! UO©Dk C,>AejB~v}Ѕ ]Ѕ;ɁbpOvr~AFNToQI;aփH'vv? E+\x?py:[N|Mg$7tߺ'A%>6`Q؋`W,X؛~SoRu DQ@0}*z 52:Ch TuJօ8,m NdBݩ]S)k:<47E\Z z 0z$PP(##F9 ƀE0nL!#:^ܐTqƧ-jUy GmnTM&>նz`[F-医wϙ@ނ0;,9rP:rWՐnc-,jVSmay62Fʽ]gǴzvd—4Aǯ&4 8Ͱ{,&> zr' ol7o;*ձStӲ rUN ǼѷM^&M3t#,ޏ^VbkUvT 5ƓDd2cAOɅҭ٣˞)0jBpl3d,7g{d@h1ƨ}a?*F{Oq ]=EQ&1lLau$k <"9&`F)30 V.uL9yŅ6 K/߉S⊶p\Ah[ƩFԕK]?G iԺiA_l9HBE+%+VYTs7tkIW)˾=ng%gÐ缮Wco9a4=Utk^a\OץzWe FsE5z^I9zR=W/&Fç ڭE\8$fv-H֔]2Ud@1\oLF +fCY2 >d nfH#ξOJzw@\\^}x|O z9bE705L;sG y"$:zb1fx -Ӄ/DvzVճ&Z^Ҫpb³2'Z{&&<cڍ1~C=vIwMCϪş }^O8*,ӿz/dY tS4QB1սopa \a-ovI* `y $1Yh;j_"2K_p+T%> o?ɬqZXZ5} ܝǛnM'[It55^.*!@E-TP+SZYPe5U!XP3^=4 O級Qђ7 NKj0E@/@Niľ= Ńe4Db? E+[c[oΤOҙI:>Ig'3鶨[Jsu B\*BBjܯ!Z [K*{:`+pV{3Q]gف U4["}Dp͑˯]&+8ѷI@fz^ Sq"*X9)RރYȆsA)J'NP9u&j@ELnYW.3~b%_+tߐSéi#L6J.17kl kgmu-%l|II,8;x`qj߯beLG ;RR&)=&x6Jg>:︩a@2+ZۅK>ɺXҭD24Z 1e|(2`\wXi]4xnS1eT[\-:Nlx,*ýԖoEdƙ}Sg$ɸ -QPd  >jPj/Ir2+\f5&"v@a+6I7XC>=gO0݃5%n8▂[ިF sAYi@!X$xwqqg7A P9^`|'?˱UŤ,PrzgțvQ,\|_ذPZ U[C Hn;mW&>Gb3{M0\> ju X M CWJ}S~>0 hiLjkx mK_x'Cr Mr=nr;$r)*Q3bIZ杧>:q7!M |GɲúoE|bR]D@)?? J Pƨ~vnZWaqX4T$aqh%B)&\ VJ1>"ʨHvy/ڭn2Mm |Q lzfVQ!:P+Jpsef"w=PRCPe\/{Vc y FvMOo959 ݤ(ᤎ2NtQqiӸOC-]tkZY(|j,w`6JJ9t]Wө*.@Sl$m '=>bś#\mkR[wdNqSTCj [5Gn$ғTBz.!mw0`-'A 8R54N9)kg1t0B`];J*Q:xoyރScS*]TQK79PttG; 7Two~zNOץs~U?gEOaוlod6ikt~Q^},&H/G2%nfW%m?}ٲ/fw|sc?y&RS726DQ{޵'ni0($M[&gV,eg6a/>_+=7/|dHV],^ea8h$IcL@;Ք(Ym(+_+K-SY. 55`58Y; ȃ#h[-g$i9wTJB"rNgڲ"F^\0Bx͌֠+tYBE !UuЅ!Z6,> 7nv"l~hzQPT&aZƺ\1%ܱB!Gdp̨tCGaі32(YF+4Ք"Sj)Xm4чz$zN G )y]E)NrȔE;%5ˎ{ƃdBJj 󒃋vig Kn)=m`0'GbNEQu 5׎" ]pn t 2K7zj5V jYrU$lrV>R͸UMkeDcA.8u&e4 PpVMJ咬×K i瘳ۼ^2pdР; zѮ[iJSVvUWVrG;Nv}mҡhC}:د?Ctd!1(Iю;Z m/5#峦INʢ]{E-WhrˤK4ߜG>:tP'9Pjҡx&x.`&ʢ]e:ċt@ g!dt(v}ttth6!Mm1T/5YtȒ$ 9}}nX;\ϥ_z)yqT!Cv ⻳E4;Eq}*cK؟-pȔ/S@ϚO)җ#Vt3QO$Ni<\G9`Q kmW{M!z⤧E8hBDv%%M)ٖ;,Ѭ!iJo8G kZ!#P3D SmEvϋ=M/ڝß TJˊZ=]Vc. A,SUٶg6 zy::;#VZlV>sPbu"fQ'-ugٌcGxQ?U1^ގA|݁1 s([eĻu-VbJra;p-RN# tF0;$ 0 %qAn_/ (_wͱM7~⩠kwm^U*~؛&tcg<]aVA쌊lv~E1f0^^&MRuUBy-*o14I^hu4J$nI*'1'6=[ovm'$).ϺP3BnbcTdg"xF@ؘVlJr@;lxgxq.Xc ^.*4g@ )J"yo=uƩ]-SKl[8+P˜yrzج/ǚ=> (g`n*߱6m'0Cbw&B ֨ΔF~DP`9mNH1:2!2f)j fKZ,RMIt(Ǧv=V׮G瑤|~?ʱl&d{^'>x2t>χcxM0vZ(䮛gT;"S(^ ח>+G&C9' HCPTt)9sǼɿ i*zhw4"*wsǩkf[#!06MZ֨u>z@Hv*^U>0sZnr,k8RVYPv:W1o,y?~Gz'ΛQ)UJn_aGdڟ|z~ tTz_S`~50ATr]?|;ԱPz_^ػqvட%aGWW[vލ zKC? 7UX ZxzO/HFc\3`864 -*IkKNS?ȿx[z@M/f\i7A_ wiOj!p#wנ*h&]}ʳ0K|{tZ.\Ho7D\˜zZl;J1F˽FQ%sd%/-ɠ#j"R"":MYomQW#PG3͚+\r(QY TqdR{u^ͻi|rY[[MgT yNms(] Mz5yTF0 r%%amQ:*۬2 ϽKN#:>+J 9 Bێ+|lKb_HǶ*q T ׈B-#vpC1MVEm.nU^̡0ޤUzJ>PA=󎰩xI bhl\N ZLK oZRuӒ]Z Vv#K'vۻ]t^Z},rBUwY[Suor0zPOw\Z_/`,k2Ye [&thĤϟr'AS I? [ amZew.ԁE<,huZE[p\{}tb49عhaHaH2bFs 5P\Ɔ}7t(1"r;׋ל'8op5B'[l { FXC{f"Ǣ Gc ~{ɚt﹜_=Z׽ F7?Y2yf 3JI-D)8Mj6ϾT!7Mv 0nfZ8&$e HQR1TsDlKvg-0J1[!0nT}U$0 }"_RF ㈁MiѤ1ϸkG%ad{U==2BmcŷGFp-?~ƠIVl=反D+j46Jd;1.ֹPI> [10f ڰի$;lK"!!hCBosk'wbׅ簉9ut:Lwaa804k$;#n}@m0 d8eMؒ#j0$Jtx)3ʠa&<\2 2I\7<uWPqLyȅX_F}[5n.~a^cB*o=qr=ˀvorGFr۽U*,IVux bHTf?:.עKpu9ﱭt\&Qր@IVjW+G5{s$etqnmGFQ9had, 0,#kEMmSvM]cFh.9 X﵏W6\~7#m6&9vl%LI9ô\|cN9nCa'n 1r.}p.v|I:ٽ'ajz)xI)f<3ǂ`GZ*% }"B&sm+aKc7IUMU5f::Jg\]?.8ͬ .gNJlN,ꄾ%w,q0s}[r VXc'0.׃25D W^0uš'rh4OE E9Xx,pO!F@"0G #BOj, S[m&Ӝ/ZXQXB3aS^hٹưYFn#\G\+oH;VCV€R.IrNƾ0Jau}"haw4 p,K4CiR\O+U.m~a;,I\(O=5]0'Qb Ю-uX%ѼK^|SۘaxNAD1` 4ýs*~]J;ǰN?%l'yXWlښg^9h(zEX<]{|ۧΟ'ΛQA>M٢ wN7P p6̏u*zz_S`~ aTr]?|6nTg ݔА;HÎ9g(t]{5sm]`x0݄*s{Pu*<g z%sRs2 @O{fV;*D9! B]E;{&WP ;7lxҡoFؿ|uz՟8ӇDcwz*`y ?8=TIPa~ro_U?Ҝp{oGFc!U3W/<8m$A_so}c]kh?A)>j a71HFI{]"?yO~+ 0 L_C )c|^ qYb%U.5z)+|-|[zj璑&a6~RX]z!L0ǖ c>PDڦomul>̧0WA$w$H6C_utZe<÷b".kuyx^k],lY:Jy"6%ɒOدH>Z(j\5lmR=XP"Dsl~̮*'qJv#^܈ |v_i/Qc_+z2gyL] edԶ d?Y_aK`/1%W̒4H!KiI~;UX%C;K`IaI,Ƀ%y$+_} Whl (UD h9)DLI]I-<3[D^a%6o.{]c_H}f?3ܣV ۞ud6spFrVnWAZaZ]Labũӝ B֛_FQjx6Y]nϋ=M/ڝßLz:hPI^Nt@GF/Z)kc*9BUGbM,^ѥSƎO;FXڎ>aQ֋_)š!'qR E>i/7P$5vg{nȶ4,JV&B&3EgY&u޻{ &8w@4#<RLւ$D虅!oocl* OB:!0[і %sڒ!Ҁ61\jLQ4^&݁~Ȩ3usP;0rW*s7ԥfI@.b3Hdn16 ?3_B"ƸR{SJ`0G~Ɏ_Ͽ[i2 *,eLPJ (:0 )˃flPH 9-3u_]3$ȅ@Y$=yG{XuHA\ Bxx"R@3xdVrVۦSU!7Ycݚ9a=ir ;C.o/ǐ}}ﴜÐ}! 9\{SrW2J% %p 8GhӟxW2?>R~|\y rUtza=|w]L'?>jH J^|3)^3ytO"'.)ێyЩ_i+}B(#SBtn-jx~Df ONm^^.$^n{ gG͏I,ܥJolq=ODQ/DD |v@@m9#XHS" SH!jRp!%aaRWMa]/ O؏4+5r.t1q:|@qY>:6Ved\6֋ ^@k<^43T0iKtg J Ad蒀_ʔ4Ӓ8 b H &C0e3 ) 9ب\ ) YO  )@vVvٛ (54(CrPvayzÏ{Hv/ _B[o0?\I~Ѭ]9u\-]$e>FPm#PvV6}_mUE~^M:J/ V.dƨ&@@D{b]KKKKפ,]+tD] *mj" d 9SIFRVHuUg׶r^}H1[@CQfk IPyֆ:(3vFnTzifm"Y2eX3Hݳ^d\MK3A@sBS阚]lP!XI@kT9Nxlw &L'xO0){!BX F¹&LcCDCA,Lĩ% D1_ۺPLk@ွ$gG4hcP(;5{7 ${'}= H_7ҳV?Tx'HfV~+9u恷/O _g KN6 p>ykwgr%*N=߹Q4;+@S!kF9 1Wcbvb7 {7?jGGkNt V9:Ps͛cr!s6f 0tQ"ا 4^c9@)v'oi/ovOhB,ր7>N;8lpi~GvnKے.ðcP):]-@WyNB| )秨uN_I8?">i'I_,H1w6ͮ$]M*1 UB fr[lWzZYϲ<׵]+Ǫ> da;`҄'J 8w eMv޼.SDW.1y]b'IzJ)S,<-2;o̩RĹV FsCt") IdB%#B,3# rHHS!AʰDLawǙ4qg gĵB#plgw\To y57Kۂwj5:nD)n mEFeXf 5Zb2kSAK 433e CLA.4"mp舤RRҒcTV!$( ZOQ31Fk]wzq,A)0 P41MgX -yEɗ!'}:TR&PM3J2$|ݰ[b{v}N 5t8#Yo/1N뛙fbuA'"4yX."\ƠQؐ9HWp 쎛c~a I}oEdyq!09ڣ c=57@%;5Z|,z2O"qf0F1<3>ᩙ!ox熢> 3zkW߰ξA S4S}<uO6Yą[=TBzR6pG͊#c.~gDH P0qyqG"ր0]J};B,Eʯ8o6O913nө6T!y<,D0&62-+|v$)UݝF~v^AR#h8^C~)4w ėj# >VV(O#{.TN:o 95+`lU.m p)q#}N3O Ry$-!ӂ4;-.(d Ȏ $x'+Nuˀ6e;@;ƛ\@#7M4ZlY6IxۏWץ.1GWn͖ =mw.1GnSu vM|AAT]Mba\j ;>#BQ_=ZgG]?%]hf/.TO8󼮁";eeq sxɜm"pLj ZF8 S9-oԑLHvBAy*?Sz1@j!#X{o-vL)uz 7~ԣ|r Y"e-5 VKpL-2R?(-dJ5hŠ͑ V#_9 ;{*K=L4 kx y>:>,]8e_[ vY)#nZhWюH5Z#]l4;"]ގJ*'D;|P]1ugj􂺲6`OD]NO8hh8fx Gb7[XdoVD&a0u>:PQ-!Q㔉">7ْ""cu#$-VK1$E]kt'fUS,~v8@<^AgJ ?} `~:9R0n{Չ(/k&R_V{^I.f$};Lolq-.RK˲Z@g e/p"Ε NY\Reb(& [|_=4r\/&o} ؼNIۅ.&fVj*`g⧳:XxL:* jrlV 'Ia2Ck3bʌF[,TL(&-Y(Z(dF$`VI#sYrgmŖ/{ [C 8J C*9@+bۓV9 ˜4O%lLe!MZi5?{Wȍo٢p$վ{&&vv1OpT3,QZj"%E*RVC^Hy{lJ[&Oo -a)!BTMm8,K8W=Z)YeSa=xeN.'X.rޣS3̆,[=ط(KG\>)#d rV٤vl֠Cs bا4B JCPʨdڧwZboK"J,{`oI]|[FP0𼏢1S,BEaN4ǣBG1TF%;N^`/&ϫI7y{y6Wu]z&rk74@)4T'mb 21ar(%eݡ3r;Rgʤy Ťvoti^rR4?`?\~vd~Jô:g}Ex\A>㫧\v`?}?\P>KD<6bo&0Y^l )/'ixX׏77E'sc]5<,& ˫lKn}`]:)Sg 5uP|4<:?-~^1ޜ˧*P Lˮ$rc3K|KgSclzy[PK"FF=1IАj:5 X8 M{!C2k둧r- ^u|\|'/YV-ocȈ0FQq2MBGR[M(O,BF<y2+N3@ 19GP+Ud 3Z}e(4h1 Qqͧ }, f꩸|8- S( !|.%H@9ں\Nğ^"'Z{f̡ UU ? @u$Q\,g ]>^˟ʨ vh4(NKYXJ^zXk4ޒ|P1~*)?,X"$I͔p 52l"A#R!l-XFGhN#P3}r[A\U);vk2Pk7%S0U6 ܙin)PlW\}ԜoXZZlzPhR-,]B8͓ ^J7ck]G$HEcmBs\IN1 qA\k`싑TB-5ë5QI-9%GheԔMPIմB9Gjb)D&­ JBM9ɝdrE!Ew#dIT dZQS6% c&K^[?Ɛ%3a-zKIQb᨜'QH$}lCu5+2\qhs|!j1% hq84')r5! Nd_ P Ͽtko0ydm]UQK8heѶz0 FӍfYvX$xIQee/G:HY9}*؍d0>VrΥRj$'иRk[cRKsweDWc FeO_ݼ`w4E4a;H۔) GM@gt~nS*SRm?gS)4@ڑN%)E+#i2oUl RI٬Ok: 4nFild}8!Ѩ44W0pJNP L= c%\N& =t^U5kIօz&co\pt-Rۆg*5}'NSRo6S`ǰz7dVcSpm!U5BUfVRL #:Ps mH!eg=aLNыz趌E]鎂~-EzZψxN^Ua &k*O! Ҡz]x'aLYE &ޡF52K̓1C㑲JNƖ<1\*q>ԙǧVh#;0"40Op#H݉yh pX"Rpg=aLm(1}TbWxRm1#39( P&z} 36RdY`VRz*̻@ x"oO uQU7۞EDdtl1!HD$PKgA$~ʬS sBQDiD{7;ZaHKv|4J+(;5Y!P92mmP.4: N-! 4iwg92BӺqѤ$ ۦgM0O ]J16QMMvV]Ćm)@TV`PHg@=;PPKn'X6(Ό9z+l'Té'фd*rOJgA48P]>Asĥ>ԇ\GE QQg>CT"JSj1R8#E 2RH3LJ+[^#rhU0-͙=+}yݳx݅3Xb>/fݻ˨FaB0Ҧ1rXKx1L&"=CkP벓Wj 2 |rBCwԾPZTa+2[oMmST|ZG~;~ t';qΐ A]{<Ȝz,8{Fdxh~7zt$H|;~%2Z#LpmwӋR{wӋog;z<6 7xei\9(O[G^mwֱu@j׾ˍw`_(  { yLnU6J%*p),譖" 9SAp=l౻s,M[}!ԫ޾KA^÷iݎ575qma)N(Rd2n!%')FH&VFrX9.шUlPTWG* D3 ;O}t$A6LU69Wjl}c]zәa #/5]|-N)CWj|͛I'GG XoH[d(&S@kN 4%3(k@WWvȿ3'"LJ(V(dݔh|Y@d`ì㑸<5dLˑv> EW (m[G^Y,F5_S[v# \)Ɉv٦dDwMFl^4sOCtc*;_[<1txo(]AֻZ ǻU,'&t\BÏYF\ѝ՛F^mkN3TGvTV'_Huϗȹ{~iyB3V j (gӷRC]t[u u!/M>+w+9 >+t'vc^ c [9ka7o/}uAcKZoކ\^wF^5h|I],Vh'Sܱ-ይ2hIw9>Gis3O7Jކ`rFuq fơsDtO [%)."o~9Ⱦe#@ "~.x3t Jȣu!2tjUAӠ1q|hPptWР8 :I8M9ϠhI~-SF]2:(8S--Ebpڠh1vґ h푗uĈĈN9pU7y \ڹ PF][o+_gl/PiQӝi_Z^l(*MӢ}34ŜhjXE׷.L- j29eq Ln7R$G0V$tDI,V@sX`lq`&g!T"$9X ATNIikH[BAb?`-xN r>H=aʆɸ()T[¢`I@~kzX"ʱ_dL Nrǹg)R@d7TB9K̠&#L)\@,sK#HZq7w8wTż[g̩E8=TRg(t ǖȍ0ZZYC~I K@ڃp&$ǛW>Zsw@h)gp[.L3q6 }HAYzpT^lJ C'DEIN0g{lߜ3찱dz#RuupQkk;֭=.p佁Ζƛc";;n2FG0(]I^0Xxubp-f*-Пk$P ^r^!<#ngVYw=:0*/jE{a2*BE@K՗*xvdd2d2n~2YKD]tA !퉢$|xO4HAJEIE:lx{ĉS.!kCx"@TU'Q ]#NJEW^4&eԪ]LNr|OT!׊kq-\ 53Ks,iDؚ+emDX)q^3i{Ӱ~a1I'(S\z9n! $5bFkO\(䠤C*DL"ۡMovTO7E ED: Ufj *MsJD9riB&㘶rP1lhwDŽ8{Н` 쏃X:FZ>߆6t>&~R:kQك._ .䑅5Ī$Fw8 \ .tZh{yv_Lj`Qv/Vs0ϩm3G< ey;'3OjmN:j?V ˙L~N }ajpo&KSb]=KR @jh?ڏ=-w?IZ,7wd,PIh ɀafPH\0|rt$OL%=p@%RSdnoh埖J]m*iTPN%ux9N\S8&L쫅CܪZZB OK1ACT з pQy湘?W)9 ȖuSVRdnN Jq:7HHwO ҩ(nv^i[ϸ#;ØԎj1 Gv1@.t@+`&~}T^1"wjVě$PEY2NMs=O}F$jB[z)Ww0= ~SB w%ȳPvMv|*ܻ666aSS:^$tiOif*4˩WƭzO1~2Ձ8bxqIH/ޕh6VW7)sfrվG[e”dz^S8;yvr6r6$R=i^ b右Í ^NԄhYM/B!(OI/y–R(ko^ɩ|IA1j*%Nv(L//j aӎ zz/_Vٓ`oR#JGXR* Ȍ|$a#]#gol}}h~,!SYt@gnPc61P̑vFwG#:ob7oՏXu$,7(24 rޭe* 3~ZqĒJ %]Slb0ëؘqr*RzJbu%H6WDbbZu55׈mw[uVwu'𠦀G6/J(葹/O#;>Ls ?zvVWZk¯~{\3? mG 3U}w?(py OZ,^t%@!ڂybo>^Jɤm3WR:̰wlR٥5L+s1'cZ!`p5=S"FMC!쭅ܚ[Q67wٴπR!5=I&s@8jebʵ>AY qg xǴK\vin.)L\n㦱_`GJŒ{F7~#?xj6Zݐ!)2ɭ$(~Jɣ >ic]X:!FIB]rN{geseb0JF|^sЫԏZ;g! g A1DSs`Dn)^(;GrCײ?RRi>HktLkR@b1z-8Z+/Obpe0{ud~9~_GIo|H 4jiYH?'m 9 B*Sxnqc-&Ȏpc-jq1J9|>/#;2ǃcQyDS:VK5$F`p ( 3z`2ڒVͯߊo>$ ἜR=XI6r_^5n?G]|nr3+??'|36?ɽ-IzV@S[صHj}0_Z*-kcAhVFѽa2NQ.;Q]] JWDW=cXycKlBw)vs'!;(G:BbB`B}cש@LI{; cn#NJMP;JEn0E5$)?XW0l- P +/k/K"a{rG佉(m\5+ID>:ác)*h#PF5ѐ[T0DM^hnY}TH)TsFj9\K 5oLYB("PxޖGP*ͿRd(Zp5#VX| ˑp@TN!grG u 1(|LД rfAj|ZK W@s(,\:?$gqa $q gVz-i#aIR!L,A@I/*$.A/ǂVwo:= &N}r:o H9 %7EwVOѐzuO}Zȵ|K4^_aU;? )h,P"sE۫rR() 7S(>b9>R**ה]Eۃ" ?o si?Vj_E>}ZZ[teE,䙛hM 4Lh/!L! tRNzK[2ѭ y&bS Ǵ0h-U1:Gv])nɌF6,䙛hMUۙ6ZjZTSe <3[ @ĞBF2q9P3C09ςbBbgXnrg9PA.01H#91>$ ='8ep5NsYɓj!ovbKA0qZg'z*mgob.ZgnI6W&BtKAQ'0;:vݒnmX376E@f=[ bL'u:턻 EdFc[MtǦXMsV)Z8Kcr dcSd,l`V4C~fPgV( !~ᓄ('5An8WZG:f~ /|:k R18B=X Y9'Vs?Mj(ԀՄ!oj~1/EU \>K``Zf>K<x]߯De<(.%f====ӥﵮ(sx;VWʐyYw+c#g}vY L!{wYbj_pz<;Λ_X(g+#6"PK(%n! "-6Z)4P&ԓA4o>\cӼU0bi0@(V E2DQX,74ApEL[GR%G&5L/]j@;+RȄE' bdBn1D3nMbCqG(6&' =N{+#kL~~{{|vFgy}V_'RJg9_l"Vީ.돤VG7k⮬-x҄Lx(ƙnѾb4ܽym`OwA:k-X0)R2@mHeBRʊ$"S8Hj&aIZRҙTE)%j J51)Rָ*US UjuӰojvܩtO-*~)ZDˉ' y_E0>-m&TT-kd9!!q)ZgDlJҷf)T;7.` *2Ç^[ye{ y%wtf08g7zfN(;s|#m1 $,foB2$/.^2%NIu˺9֝uAmukpS*Q$"w,Lwn7Tys2W2k\#*g2.bVj 3Ћ- *ztO: Oγb6<@Ni5 iQB 3(TERA<,8nln-8A87Jj<OoP?Ie=l< FUpll[gD5l21p8øouG-ayM]Fm*6M'Ĵe;Sf4Ƶ39(h}7_~x)aLS牥֤n݋{g)Q8/v{%MnBFfDp;ayHӓ4KruyXyy$\iѓDs݃9,q4Ǔ"$b)vXTm S\T\D FZ(ތwBq4ZW&'5R^bMaNT Tr42jǴjQQl2:X M;M3EkYYƨK\ےSJ^lkN+Y~,(f7 /nnޱx 0}K,,]DDhdrΛ_n.:Io2-nÌ|'}dwaozђ?χCupoKp$Lٷ`$kEc83;)!N '(4f/SQ3n_` ˪sۗXmnOxy(`_VXQȹ@襀ovށ41GjdBhF#D %0D aUCo*i5Qh<vatItdLX"@;r&4Ʊ c 3.%EG#Vc$) /$ǡT1XJʟG `Thi#nU4GqH$J).8t5>ZE Y1qk-!cIHyr rlgyu diSuf7>c9G5_7/}ޚ/GTS.3ЇXSX@taW띩oCGDخ=8Τ޺=96[|pJo޼N8 0lB7Ì֍GilE2K7%UF~*vpx&{Ě֠T)q$@id@&eyO_ (JQ*NzO %@X^i_̚e$FD1!O 2N(?`i =|bA9ہ ӌ:V"a0KGL&2/oNg\@jxy"3ᡳ<Gg,̯0->{kݏV~(nKP:F$B&BE؆ۘ0~Bٿ{ujU&T9b:$A{0Z C8 #DmYlDՈPE6a4!N5gW>Uk%ۚ+ݘ;c Q>|C%vzX2޸|/PqSka*&$A4VPR# x,RPu}!{R91u[و ɩx+ggX-I+zͧǪC4ΧH2+{c*^ nLt.2ޜ_ϝ)pu% Ȁ3yaA##cf8ZPOf>Z.E:SXN~`!<+R( ]I۟fp lZQgkOW)%l䓷eO-PdyZfwc3t`vojDz]+Ѯ̞civqzV'ϰ굝>޷٬mjg'slD9ZɅA`*wMLf9R_a7?\ [胉~6#V20lkߺ%-v vsh ]qCpR[ Wa'btCɊehYI IRdl[CmY;Չg3p4sq&7o孍]R^mcsGP׽<izٛNfa*V@%VܿM/LC-8ykfgsPx:MS/. ޹ IrWP.Wuˑ-S #9\ zS&'X;Ăd(u>3s̥rEtpL;㯦}h>G} ܖލ5tDR+WD#tG,Ҍ иA(bmTF!)6:T(y]t u;ˑ!aWRfvS+zwGl0<uo2'']'GS'Lffus|Sέe[HU{igaaջ D($RxOksira?_/r6^tM(w}Io&y4/^:.~juck}y/_p `|._] ν|9c~J+NY^| 2At&; ]PK41L*"T,QqZGp[K&a ?ީ 4˥yI%k` V8.,\,+q "DudRdf[1=B!. xˍ-uejMaraݯ l#-A; !`{}ɢQ$bYHrw{"%T,L[ss?>|]_#puVaox)B]fњFԾsO○Ǎs j>~bacv(}>NC{:׆ n<ۍ(an.L666ɉbϙR aDŽ2̲Y+P"Q,ji#r{󝜒B"qkJڡEïkJ4&{CޥA#q%ȸ⧲X,R%)j8D%2McJt+ 5AQb`N03pr$ɔc34 !DU$|N&8⣮&5OHFD<<3h $OI#(8HgRFmc,2ErR98x[s>]^]1viB0C!Eְ_{n6o=bie rswHTg Y탅) ]b*c7uouiu}zoxg}F}~)4p[ hM#L'^2aǦ.N? ȩq4rϟ_y:n7k}o@ fpQܦ%f  ~s+G_&/i ,> ʿo P.iKl<>)WCbQ'ք7yIS,y56&5'ڕZ(=-;1M4bf/ 縮S &@d#*ߒGAEȓ퇿(Ph.Ta-X:O+,@4^l^1bÝ tf%حxH ŗ $SCIK$/fn4fI׻Lf&բ."^FL TY@O/o?5xHx {{unLIUaj>wcVhP&1/ jdaZ7/N nT4$x!$7 ztn1;)1qXY.E& )7̈֒kC9QBUIrDMz+gk;x9j̻i遅[ I,MLjDrFm64%)J,4#*rIPX’ADd UwmY SKZnez?]rVHwq$'[ jwxnG9\绅"Zwk@Bs )usM°jPwh歁[0u[hLI,;_ 9nu#2 A }ve}KJU.v FvkBBsݓ)QSĈ ZAe?Oq2߮}࠘pk3Bw#k*G=QzzzzwU/Z%TBn{{N'<(ʗv^*-ל;[di1?zRt&+ͫw]:g]^V1Z6E(R4('TBJkCO3ҬQPw|P*?JK9oAʏ>@?$&}rë_T]%>0}`8㌟ƒ2 Z,?@4޶w;v7d`Sf&B וkzsb0bj"F=1Kbc$51!87v͋_nNPQ~BT8.$Qތfāa5iYZ⚖tDhY3lf E:,J _vc‡_B[4QpE$[5rxu)}QuRe8F !S ub(ށ\+0K4"4[R%jaX7g ?sl/8G7*吠PUtsJbVS!Lgb5OTZ̲$A`HԴykx.aJ܋T7P`@͉({MTMGWUO%T=@Rqlɹlvr7bd)sI0Lm3fnD8%uG҃?R ΨE>PX?Z?Lϐ&gRiF҆Ci8,=c2cJ 627(!&aS%bo}i܁Ƭk.sًD,K$7V4 40OC,A9f>bĈhO0yWn*N׍b//>Uv퇎R]?}-~-G`:}~Q:TURee5[5 in-TM>[";/?)#FrRe6RUPYvaZAhjS7Be4Y4OSLDp[E6@' 864RD!!OM>X,a•-\P2ʔcɂR\`JT%BȧkO@&a%S LM7yFȥ[I(Jd3O#7@PsKQ (in1PJ4$QyM3LGqqFv7_ݠ; Y..ǻ( pxdzH^jwGH୳pxnĤ&JvBS_dۂ2}fz|m~\+`iMI *6ը1nCÈ Q3>0#Q(Uʰˁzm S`rVpT+![E E Ԉa X.KR{BR`pY1niFm\W/ŷxi#CMdxF>ޝkHj/U/ t>j.foK$65x5/n!Vڵ>Bg%}~`y}rdKAG˧^#IZ.7(!0!`yaST%$٧I'>RZaJp0}R^/Ip Po }4D̅K[[M7FxP WW82pe.jB/HzBU.@]Q+S%kz\PBZ"%/|tFh% #f#yty'HqMhXW^X ]ne0bzһ2`XfajFbrMJ$+M<3/W 7eMpKCj6hqZ %vbD0BDtj=B'z se;\oW Q?v= KEB3. {.怂]cL_<ZVٷ$> H3!Nj1NdV9F 8é8w5Nle *CVu Aw"6-4¢yx6'OLZ8C9} -.aKB|E DÅ}/W cr.kB/GXrQTAΫo@ qK6v'>2#㨙gЅ/KSbƉ%K>J})AqOksP3ƕG}y=r[Ўozq؟~m]:Bt/$c an<~(5JE:}(ԕcO%:)T35 cAf_-[(J̶'auu2JL}Ԟz،hI̹B>/IqVޏ4S$2z8k n禷2CRꩺ}P ɣ 0Zbu;HAՄWA-KG>Dyd/UrR"HgHfg9{c1Fopĩ5n$n'db<'UWm>_p=^ATh3u-#EbQ˟~ԉoJl##h]|&QQ,DEr&4ZgB!׋5֟0x?#$ҙ= Ƒ 2ua bH̉z~]~=BXj,[@Ծ8˚d uy ղ')f>o$1$T‚Iz(Ս R}0?Y ŋ˶DK!rixe-#8lBQc#KzE D,Z4<Ls5%xDN}h58lwع[j24= 9A#2,14(VQǬ aox,JnzzJ71hηF:S),0~<xC̘R>ӻӝrk3vT7%4v?ޞz0ii>ic(M gwp7r/F<Mf$u7XY÷ ?Z-uE7Gtv.  ?RE }Ɛihƕl^j0)K8ow{DxZ|xnS ^XKk% "g)! OIZ3v &:96c{qpT%Tk̓,'Q =o'*{JR~)7iD- Lٛc@PT&Câ򻃕 EJ7R'.1S#P8w/U U| p?Br;n'Q#`95q!o_ݢ``G$Y/mꪂӝ@(y栫V=:C|Q…ĺ{7n^ LK ' ǓHPs${9?0bCO4c2Zo,84{ ?.^R!]Z^]a!DJs T"5%EI㘦/߫Dr^C XhHE}>%=huNz<_$NI-~&OOb P!J#2 _RX*1XĆ!:%+k i[|dPvԦf޲wo=I_יT% h=eǀ~+:}x &߁of!m"lKd9,C},fqzKUoHe!371׼57Ƅ^?]?3/y1K{ljk J0Gz6;sHfzAΟ|KAv?*=\">IH9p[s.Ɋs@+M|;)b̋9:!g;bzv۞KR\zr[E}a&|G;y`L$zv$ mϑkg Ji `k c"&.!8X.um;2>ĩD <ZKB{)wHR,0(@|!l8`ЀFN>@툾Ahx敍\UQİ.0@kO:h͎vԳ4tz؀\p;)[D༛"uԫIuG2]7>́>[X˶eeCE?K hcH7Zٗ͢i /<`(]`=40™S]O?hFw d?0n؂)w)6eNzVU{ hE\·Y: *:lG21GCDPśP ,jraa 9R;[;&VP\; v:Y<[*: P_$085X2;n$tbT1tkv)G 1[/IIHHȉO >Bǁ 4oiNYf*ki\R,oyOʫQ7' TG=+,xa1r?m& CttB'#Lw]K,dC8IAT:cW9!]oA׷>a"^N <Gz&mUQrXW]l(rP4SgH,f*DJ{xEQh9{7O0!1elUDhXcTETB# {Jgtz@hߩG5^4^奐\ny _>2]@t]V/fÉ#q(DC( !bA 8CQD2[s65{KY\j9~{7D J˙B/EMO)OŽ7?~z FϥRY~sLHxaㅁ:^ИR"`L\X )".A^BC m!+ -Γt,ǭ"1;.oϹO"քD^(.5|˟^ٗ@3W7rۻuFcO[7\a;"PwR4d, 1F)1\abFQ881i3ԨvWSMW))kE<`] zlA.| Fy-gF\}^4P1ެPĵ}[OrqYE{t~}uAA*$-]{0yh>t Aŋz5]Ε?Ipkvns&9]=977גm[sevW4,}̗#%zIhG$fޯ'9c9c`sDDP$IסHyapO"BRy?!j   l?^_;~"-,&^ rV0kpϋ\Hv}T)4zqxD{C:h |jP p8Y8HƓ( E| ˋ&/O3"%3rtT^ad5S]T57=[9Zr0Ҙn<'Y@g9̽]'=>D>3H4g&02(Q0 a`Ą{یu*j'*ŀHpT$G*~T@*8 Lh")LHB"NRY`u =5~➳+Q+Ft W\O#Ղ^/qemoIFVG,F<Wu4EDR\,W ,dA6YYCvL4[$ҏ}"rD}5b ag+\? ٺhΗI@ 2GL`hnl<0*IG37c3u.K?{_5-wadÊZ@WƋ0 t&Qb.w`(oIց%( (a<* ""iBˣPGM"H ,$aےl4m5rL[З8*h g^M zTms1 m'm-/#!mX )%ᶍ0q3FơYQ]6QMTՁNϳb:R1TlV`$ӮѺ3oAzk?x_ng*`7 8ڗٺ܇KtSuEofC‹/n,V=z9 }$  eiqA|$Cggnm+y f:Z ]F0jf_2ru\\̶;*|Ͻαz`JD@ L)!䑀29(E\',N"IS y %y$Nn(3!C35h=? PF]Ggx hT`GvDIl@1Rñ^gp鍳`S{k `Lrʙ%4Z=xnBaX'@rDD;їNj UX8p~?]6Ao5G:{`tMF{n'jH4F fNXXW#^(ƻ?_lCd׹}bTz?z~zBRrxb!\ M]ݤ)cx{R&z},ե{jR=Xmm8*7 qQ&bͷN|!Ά9rZ.}a`ێ&wE"NqlMoUm BNBN@_Zq-QKd\i7Q$HA_DBѼm~lx?5ߓWZ; |1oqTw>ᔈ8NGz (>;Gc'1etzPxkex0ګM-OP1{C=h=ܝ]¹lNǫl|f#)-h]t<80(r7 Hn㠡{3MZT$qA@f|iF\su) J}M2<&-VTY'lѝ蠘ix/FI/So2?kWzūc0Z}P.5p†b}d1R>JE.%5r@E)n,%BGX.ӜD3S;ẻ@U{ ?RZ>DR=J&3>FlzH 1"~R$.jrI?5+(-Liܵ5qۿԭܡ u_nRvEqt Z.%[IArC.g -] rszݍF7=-wv7'/3$uȉ(AyZQb%jAcEK0@#8 LVQF\,v`Ji֣!s kcM)i]6C{okpZ>/UOr`RfNY FQl]LJ YgB7h7Kg2Obr粠eA,>,$zUK(vrd4nzj_j/S%IӼP2kOUOV4K꨽ZUwnK??w;Fya>h4l \4Į ?٪^zp-#OÇpO|OQd}~#L#%v[?Г#Zmbϥ؃#u )+#P{8YM_B`⁛͜8w3{w(o׀D-Wnm*+*R^ xeu@>V_'|FpCeWcCĭC^(rB ]RDS)~ BNk,LaZ2cBOcşCb.L8e *_Sjb_ByJP3WJjT8k*LTC\}/ﳔRRɗ˔J##b5f3^"{Pe+B2n>,~۟~j4IA#svG˨:CzNsi4ܟH9HJmʝ2T ;M9Y 9睧7dYY!s;!(YZCybY7Z"GќH N:u/1h֪R3qр,oa$QX^_t9 X &-gvu毕b14e=;2Ś8Q8s!TkKfNY@A6Ap@xFk\:kR}!+P'5ͣ*!Z3?K7pG.EV֣/+fǗ4v5Ot{_5n} _:p aBik1]ľo߽B{Z~DuWv7}fo3W~Ps-3kǧ rEeDR`jKwaLrfG&[e&X3/vz|֎\(Bzu1ڰ|Yk, }$zQw(aKW?Zn`c76w>-nK!8utGXQl[/Qx%L-b_^D7]mq~5z/}uʤ WV4`/g*&w/ծo/~ƍ>\c.VT9-Uf)! j?._.L&f&u-Tѩ#jԃq_ѩ#zޖfՃ(3 @$ ux/e0`2kPA\t&uzdJ3fΘ2 1+ :,sM"IbY,w~78^f7W't;Msu\fJEeB"ieI9#PfEBl.1OqxŸlyC8-0mS!Ğٖ>5{ "M ;Kʴ%zZ NGk1 /GRZr>8|+nA=go0-8i]zy<,婝"K6]u#]_;EA`%K'!B/rޟwS1>s' 6/iXO>Aȡkl-ֆl!Vlyn[ඖlg?׌mCغL: J38F12zW"P+ퟘYԩKPh:#ŇFcDrF`L|9& H}8LtX8* |=h)ħd{gkl$@y,Y?"h Eg}P2gι$˦L~=' hn/!7H&:.W@2E:_~L*f&F9ܝ _*ﮂMכ?]>.i="St%ǟZrˀ+4%H-ŕ΄9'5N|Y͝ u&ThJ7$RG_W-xH9f#a&q+и3u:9sT#Ǎ1ba#/Bs&79a:KفÃGydUWI| v 䄈IYp"D8xƛ`s=K# (YکL{m,4#OLA)c['i 3-$ &@Ʊ~!6Ei;8]9 JftH^/]a)hq1ų#ly{i?OuEj\g_d MH\MHrT'FќG,]N:@P\:, nluAkl]6˛??$; Yl\=eW$:@?UN P}u]Ϣ|%rְD*82ܘl"8Uo/þO|ʎD88V3kjj"Jy*QșwuE"Op\ّN+´# n9o._O^:@LPL(&k؈I*d+ +᢭s QOP+:r56Fce"zcV % b` LeG4j[NLcI+/*cGeP*1Y:|+Y2Z$V0$'*)-R! EFAzSY "&:q,(B_YFI\q$hvG՝/~4L'5KH j~i7rG VVP# A/$aR24I%r8ȴAijbq8R*('$eFG5}΋Fg_՝ΏgqKyxG?)}dv QTwu6ݡ`>B)ATp'I%`T)znPpdh0CdHkP^p]҆qd(/:%QY0(#x; p*hya)DF|SJ[4\XkJɀ}ZfGx@ܑ|x9.}\iK;%Es+p%w;M~ZDP\Hb~+`wH9HƽoK-t7%8;GUF!L$,u=B)˜0ej+L:HKƾ xVLLpPL;,G&&JsQjzIWB z/xƒ,(/#+QJqIF5W?沋p/$M|~~75n0F!oFs"U3 yÇB>b˻ޒvd$ffUI^..W:{JCd^UȂ' y#9K]TB75zqwwߏ.9#[5nbV𚭬'ɖ[% * ѓ,ta0&tͳ]\2ƙ't79^"Z7^=VN/t>G/U2'32 7{hLFhL皡CU)=R~0rJH)f`b'^ޗs'?}2zYaRX([6ѓOE0L^?35LzH;*8<8 9[D^<7D#,M@ (sȥ18S5Cr]7^5,*+WZCL4=Ɵ9송/> jL4Z5kh6.D0!Fc&1Vd03-8DX*^@acf[.ʨpA3d9m[ԲHQYo uU4HOI0pW2z}j諸жT_#W  0k%NFaѪԠphS,8R8ͥj`ՋC0!dS`NWͣ2@IyPA:kk'&= Q۳o! ߿Zv:|+ݥ݉y r}-Kk.8ESE8{?^pç|Ex]a{1ړhH[ FZ A 8uDۚ)"j)`(PskEMl 4T[h Fo<&gmHr HuwU /  ~L$J(;CUHkDҐ3$%K԰^]*92]upNI*ȼ$Vr"xp戤"XrQI2;Psd7ƇH6xu1:u4 _F _䬌}xy^Nu`i>b^A\7.%^2>KI_0 ߕ2xz9{?]UOe̠]S]ξh9~L*؊H)1~r;buꎚ{,||xɯTf䌴&R>ϱ+ж6m=K&aC r 8NX@=,x,(,y%ڴ$/d݁[s~F!`w'~IqKPcCqX&wFN-@WFH@v :V[R}M ղx 6_PeF@c-RCЉ|rC1!ϳv¶lN~s$dqp7\L `Ֆ)\Ei[ˬC<*DL*$8Q K'LA(B)ljA0x-d 9uD^#\R"d!|2"lP| -v;Sr$9E#D%E!#":hBT!%^ej,[zynJ-S[,1^t 4 T$aDE]=DκEdZKqKV29Aa\f4Bhn8JtD!sgV9di{w"ue*WMR]]Χd^W)o>A>>_}T,V;GwB~ ͊C ] $c#}7Y-Ȏo>䛬򤨠k݋|#o8'zͽ[_nL]?&eX.Fgn>\ɩjXi,ÍٔK_XʩӇlePe\mh/d-)ZitήL. 1{wis@(eufb g9nbJmNvL;+R˛tS6}Eo+@띲_VYҩ]D#K:U^ܞtA,JKwj曹K÷L8`+1K4wƬhɗ=%>cߍS5u[).|013 +X".j,A8=t==tHhbNXR.thhvIG&v$tv?~s$ c*QD ui!nۡ߆Aw<3҂o&iaxIx|oCJf]_(iʧ);2^Ur{o]ծgs_Z VC?ꯀ;,!)¬EVOe6t\P(WNgR3=?}ve'sG,0PY$6r2M^P0A[kĚ۞4St-Cs,wO;5bКw<}fYcA?4‚qgs$>#nHvčՄ `:U]Q<=^6\`;/%gxŪZ( N/(Ďd#[<[W103b\U`-jn]Bn7݈Ek ^5GˏExvT7PxZ(^brCz09jg-Y`z`AA:di_[ֹ5^5y-G.xD1ԍ*iq' 1gUSf "5еVș[BD<)b7B|lyk yjJ{M'g40;u_F<^/ @fPh2ED1n6* iaϣH!6%amH&_-HhY9wFeuf8Vٙ270f֠>*}>:"A~̭,F! ͐1Usr4zGXz"'<*9Q JNxT ׯwtpX3.: &"AP(I": 1 zG[2WF3*ph.̈́# oT\QqF]4;cil8 E F$8\IքV;1Y2z27ھ4}l52e;Mή{tNb!8V;!' g"u#(5Yg^c7"W_M{#=d[6Or}2s5MiEC-}$^ڇC"jЁ4! frR;Yr& dGP+#1ZoXoKR#=[SiΞ$ ƶ6  Kcf =i@*U $q: $3I$ sֈ=x+2Mܚ32!9$rޤ!O҆ %F+'RwhR搄h2RWR> _FessWb?݋WיmœUoKaIY/7ceҫi~,g'}0VW/d(3槫SrW}ܜ&T|wvh¢2syrfi฼*MѠs>W -UZ\a'm1eK顰L+cלv՚ mF^% Z.qIN<2v/x$tG Gl.x$X!xV0Zq#CÐ;B<`l,x ڮXSp;.a{+WX- Vxm\@BXM2)"p[dd rxss a-2߆GwWYFHX~Y }[n |Lt˞Q(}byyдf85"BX&A1h؄Ԑ3|&rh7a-(OkonWjf3ɟaWڰ&3H Ev`hS G?c tH9%}pv-XPr62_ & fWiNů9֥@0WCeI:j/7x~CU`o~} 06e#sC#)$E %G7Tj'4dGg.xr0? Kg瓁!FW?՛_^DzrqTsMaH޿g#.hk׿ǽ8vY~x eLPNS,rT <yr҂ȈFт+$b%c͍*ri5MPoh#( V%kcƚwzN{ $/ p ?K %~oX6b{$T*MW.ֵ@>[Ap07]@|җZ;Օ=.o |C QjuS n;r=!.r>9'ּcqC!j/7xʋVA Jٞͼh;)5ql!ˉ26ZP8."-NbI?7Cbȿ՘#gum\?[7.+muLiR&r$J7<899hsY%ˢBk`Nցu}DC:BmMqM!dY%g}q#r`م(ȗ#flB@b޵5q#9[pO7:]W٧rFT$K/03΋FwnRCQo !Q.S#Zù+G4DZ%v`rZYzU 8am>^g{Cq=-~W,~{քuxj\OדfqT}HF5RUԉ͏f~1bΎT:}Sf~Xʬ,Nrϥɭp']Y5Ч˵Lg/%Y'gfr6tmř{9{zQ97]e4sХf7Wq Zbkc=ÅB]#:KrjU8:nA-igցB'iV@E-`Zi_`ޟӨw@X`KRuWhG7s ,a$is%9j}}r Nd \hSE3cf +xaQ9]EG^U {ᄎ[_lbrܲk~-]x1,ٝBg0zic3 ?2[ڼܹ$($7 솿X~, ^نͿֱ<2sE5qU<<Ջ;ymYb.hmLDJӅtow&AUen\DXԅغFF!pl Ё+ےd??tmqL/{ @YSm}s;'k3E/(3 ; dGq#m i#'U繴-Gd{TuehV={#=Qåᦊ`xF!]DڽsTi6C 9 |4OZF$bd<*ц2CV%e(3tFY DTZ-c%ŕB#l0%ePV{MPt"P]%0/0F-@F8uq$[RU_6QhD M!Wx 3Oni$*!p7MVΒzT%yAB,d6޻ _)8ɝF8mYM~^TCy{fQju^gZ3-׭;gr0dZ6dTMs3z;HvmI9DaZgvjd60rV$M4ݩ XI$nӓf9jk9[=9yIU.Ŭ+U Ü*kFVP+qU=\N) 4($Ch;]DuQ!=(?SqJTD nxȸ] mn+ZV>hU0ZZw+InA+*QsY5wV Jg+f^:xLAF(8# 98Y3tf1sjVA1iG#GqDT8sY0 v~dJkܻל0eg}#1q T HS,M&1xi4V[ka., \ ;"2jw]]#$$ tp \%{h@"%Xd*\X`0x-b!2EiD5rD#AW3W֛;8ոj{źF+FjQe:],MeDz(`q{.@d4zҚ'\̖v/^l- &;+f{p"rb^e1,cV3 +](Z3$ 뀎b2:jw]E13u + o,†Qm(%R"+-)UP NsP2`wFMxܨ !pQ}Fj*¢ūXA UNI0qU`jiC!,ӂpR]җ}t`>x~`u/+CЅ9k4X,`o/*v>Z!e#>=! =7) *k؇5lךShXJ3Ԓ8+#_6΋8P2 6Hn4}(+ǁA`S]*#qgMNygqسUW_[_)dI1AqX@'Vx3d\`@H 8@,%pVCྃY嶗|C'Ntʺ;@} kIcK{ޝBKŀ{#7L)]( m"`gOg):Y#Mvo^ T024| ڢu8j~o_[qR 萓4yIU{Ltd bj!=家bvS-+gNa%#b>|>D*E%I_K"RIe{7&A$uf{ֽ0k*ޮNڐ چCN-kOrT_>ܷcvj1+i&9N؇}"յ~fՒ5蜹{35P&uӬ7sSWΏ ~iݺs*#ɫM |L9$ce1:אc%6[qmz9{<9kn5!PxD,vDJK$~OȡZT7>8<87 6CQ qh9syIEXdRl;OܱIxjX=;5@.DwLhfEӥpVi/{L/&:bu؋dKdMvFILc_bEѥ7vf(B޸HNpz^Ƹ"ɉlC|(nq){ C63ۅ!/*݋]$Jvz|vc. dʫqxlE2"p4gmtI>E#^/ kAįZɆlG^ / */K(ewfq2:2\FYERY%ӢYPfPv^5' $Y8IG C頌i^tMѬP$mA!Xk8:*I,>m(_3+ * vY*cp6tj3Q XMG[ZV% Vc\ݓ^x珚KYyDnn (w19q=tZr vm Dj%ؐwmDyBpa TƉ7hZ2 gp Q3ղaDt A Ш/D(u$7kzoHɬPKe4Nυ5!8e*;!0D1Uzm`Hλi,Q,$51a Z´+ED&.0hYBV;DpxGM2Do A)dI1Ar9X:`;91,=g!x*9 9 `c.JxQ.[kV*@w7J{3ׇYy=+:bdb&3yr7g@J To AJk509AGJÂ' 9pi\AAm-F]{:c6#~My&_mJ )2ː.)ayHSH2JA=͘d+Ġ9>'A\2FGKr`xe.W͛*&E1Vgu?FɛҔYrRo\⫻YM?ʫdoMkw bg)a@w:.o%u4k]j1CFݥֈm]lHn{J@;aQ&/Iu(vd M.w0r[}{a@ngG(;E& [r&eSB?ftY̻btbMtۣN7-Q!atkBDlg-ۈn[,ML'Dv@A-/mڰ#7(w79DX N=n\)w_0a!GnQ6Eճ_}"9DX N=n\Q:mX[ѭ 9r)I*kR0btbMt;K&[t0a!GnQ6X^!گW~ QԵ݋ 5_%دl讝C(R~EYUOw;JZGmz CTB~JRUOun_~UCMO=+@KY;zйOh컥}MOPhWvtΚR֪'YSwvE=dgM=r;kGYӔu֠wI;ZҤ<l ܵHX'#Rk5zOC;kGY 6Ei:k auY;kmzƺs5KwZ Zg-챦6=YsBY;#k5ZbwZ iAqY;kmzuF넾v5ʹ^g**ZYkP1ǝ:}g(:kP0u4tY;kzW{58;k}gUOPwN^g ֆ'p֪'HΚdYkX MXY;kzTI&ZYk!IA6=QmZ8t<'D0cooiqsU 0Y|j4xE*0YRCKӂp™R6)NƣdyN^l8x9Ē>saߌ&U)fA%ߗ{~!/ďj| V}޸rV^ZyʤE?@N: j˳O3.^jl o>O&~u~̫:y6XI}L):XdWa.97 <ԇ%){ 23l>g*n!<*癇,4ѪNxA޿E`;V/m­"KB%y)˰M3Ix 8_͔Xa i B d&/& N1ʼnX UӨ| EAP{PL?Lp;'0|RmfEp۴Q݄3Zs+mx!2JA=͘dq( >^Qt}>(n%X}Llp fx?S͊-@6'Zb:\4Ґ;8O%iQZTv6xKa3T C5kP uX)AXc< urδ!Y9P t āQQ$`J1c;kk a8 X9ũLwkh" r1 jiͰPb}řVCϴd:AƒR4XĞj3;2D,aEBHD!DP{I'$qր ^"K 5TYJ2#!D5!8 JFW6gNI%bcp{j0PX$#!,v)v@egxQjN^($-cƙu)n!CBҺ{"CUz1*7cS|Yv[/WYxrBtytPdk[3*S|eT%Pn}A.oX\i5nioBD /yӋAYt*|o8[?glfdisK~O&p|*[RsB!yTA1Lj|;7)\,56qۭʻC))0V7p# C=ɤk<[.U?[l ɩ.qC#$s鼰0~s3$/̉Ί[L?" *}u/Ғ h\4˦nYwb$ hWVd?&Ӊ߰`l5*6+R%yͧ7@\5/*k$y1+>8TNԚ-a%ߧvI <#[Zco-K =M977}}OwH(‘$5!6Yť^CH\V "(C̺<(g_ϡV-c8 jq?W nTbaʰa2AcǻAs;'0IN ɸ 0&e$iyNEscx=BʻЬZۅ =Q<ڰzPmC4-{=8T vj!Y3AeBX&zB\/!<ɛ|q0yWvպDah\ெwɺ1L^G12gڷ#5 B3r0)[>H6|.(䯐qZg.[Hdz@W[v$I8o%ѓE V?t1\?\*XEtiAFi#7 ~~*ECmNJ91f|3S(q}e) =Fe@5o֙e,lQזRݖHƉ\ ʦgA(0fR6ʿ特ٰ rd RW As/&P:_Ojt6Qv?8 1:KN lFղ-&y_@,M66>(q xm~&a8ЅM1ӧ[V͸Fv([܏ݫl!ŝ(Qpݿ}nI(B>C FiG熣VŁZіϖ&s4 {#Ps yNz΄SVp3I eXzH?^|m|y34(Jt+q%Y[VObTӜ:*euYk5ro¿sinՒz#xԸ9^hk.y|{hvEa!^hgԅRL5H~MOy)C&>rKk:/l9'<ٗ`7ԍf0׮ɠRre7|a"SwG#=fj#ltc=խ/x7/Җ\ӻ{n]w->,FKm4BZzT]#Tߩ+ڮS8] zR8aJնfmYV&Gg@SUޤT#辸dٻ8n$W|؝6đվx5Oax~`8!_nըI-Y/H_*{ _Ķ=RPm1}H0Op0,T1DΒE(cZ*[@*f^^k»w\,82 ژϤv  Zf"!t"jҌ}HP}6gHuEHCs!wdwi^Ex-5dNg+5`2)uH~[5 )9k'<ܳTV?FLut:(ۯ|$S3#PJ;B`!DklJ2w 'n:(jD[.m+rOՂJB^C.лtjŻMHV ZWnm23pʾ{:?è [{xaq lЙ}ZR `NŘ8#pa69^WHv=D_)GݺHTNܧv=PcJ$q߹+We׫Qբ>#ӻG~ߒ1NOH:RasPGPՕjpSCv %&e= SңV4?ۨpPe)IZ]rl5xBRɑOC >jg~9 6z]&פ$ǰ}V)lsFgx85)H!TC*)Qr6 *4 !q2=>HAZ@مĚH$Ҽ"mO=yUgv!. s/SVu9L] ~e\د_HP!jm er{^GVAF3CPy{egZmɃ l Șgۈ/D2"pȈ@IPV h6caj#]* 읺![-S;1Zo?Z0gzP Hm79}h6XIiT?-f6Kӥ7`i,8 HؓMEe<y-mFh@!8Y{(cES"!}8m7wnY>wWEoVL~b6ῬJ>2c_MM[YQ2NP\ bԹ镡{ (|`IE2 owRP>(5l['IjUFRw2J|'%Z!zyǍAI 晉*!*G=CɆ;jw$A3&,ޣI虓p#(KRa`ւ$ FH$gU:T@T ȍ[ (sS~5yi{#Xp^T5y5ZVؘVWY:@Gi;B ֺJ笴rywf\ ڵe<ٲɒ5 S({V_9M*g 5^3ǫeFчX@J!^ӣCy=(GM>~j|o d{2PXԭUIGy0{sQC#IҗU#|r`!$f1"0 D!VK0cr BV,%XUn~nw\rcUW5Ƹ7(_t}R|-P~qoP~y*hψB} _WӚM%\RZ.)cJaR]F\פRE/Fހɪ#]R2HlR3?56&po[y[w?侒!}`X8'RwDRJtfW<%Iqhf["*:1Ioؤ$a=A!.ͫd"C 5$NI:aMv\grywlzZ-]m0/61M.laZ!t=<ٌGN#vqߜL̈́{[* _첁Mէ껓Vӯ,rt4i?rBj).9Yx?OC駷0٤]sW~χjǯ''g7NZxo@,(JINVI<\H%%J8LD&J>;Kbm~"Lɉ:)]+nxdJ#>PؕȂF.cR&6;QrwĎ2IЈx4,AOrb!YD 2Dž$=1y(.'}QU9qwyՐka 8e)ڀȓ MAC̘ 3 v$}2n*lrii/SnοWhZ5 ] 8DFٌr3jFLDq!L@zFPpIU=BM(嫙:_7Bz^Bm+ʺt8&1[oADeR"p =u:g4x\D<-?2 :JBlnt"M'i[G[: -yϓWK?"=s^159u HW늍oׅ%wp/N%Vu炄V<kpऍyByoҥDMq7WMUѭie<ʦyMm"btz/Q Z sNihGd:s# U!$Kj_TuUT5֌KɮT\9g@'! {'"DE0 }@Xb/9P MSIT/V­KٜG%2IqIEҵRs֋kH*{26fضͨHwOjΌ@?K>l@Y+Ԩ\ HifF=VI2*Z%J Wb."IFG[)Β"m.~NRM4Q/ .|ۀi&5VY!_o^v?݆{:sM{ o هg” 3gOzMPz][:m t("yOV{0:$q:)W̱>RN/K/h%|=rI9!lR~+˸jogĪRƶНFJ=>*T =ZC GPa6$)_uG(8 IPd$[zqCOaԐ {6W I:LՖb[[nFoY1fT|FnV)Lt!(tU؆;Շ?G>̇.ІMG"c@@ a; Kn(FeEF ZWFf+R㪵B*'Y`I]MMIzN#m )*JcA'5%zOCD܏κȋDNi^i,$N]Ӥ_*e{dvF}o]ǗW6KBՠ.O^_){~}+4~:;49ȯL!Aˊ.\law ¥|֌@PzHPj­A*1YCtV(hZt}FY?lE-ӔHA"i41@w</E;xHtM#G+ n&Z}doZIR*$ωK䍔Cr(WS.IX# |KE i}l4HKdeRʸTC Z`\pA`'Օ,Uin$zq57j,?A37PFӹ=o{>?h.YlĕWFmzLٱ&ƈP:S4  e_rFq0n\vT>S(~(:J%&qjEB t;7>}KσebfZ-d^1 /-&$Oۤ9Js}cxjd!RxT$Ǐ﮶3 wQǣ|yuxq~~U?; >@=Zfqg`?lSiZ6W'/TQ"JPzP;gKK% ?~l?UYj\K{-^άvQ jL/>/DJq'U +&ggږ'/a4H ӘqBWcE2d -͖-][%X1ft͵z#:c3wh?EVaiT}`%ƱDqzG:1v A{DA hV{}$pI/X]rw7AkJ6QU%~&.翝&l 9~Cw65۾ۮ-B) 8՜W9kOs)9'hGcydȉL*Bxԑ@Db(C/.նm5#ːls ȮrdvSzu-37FJB>fIe.R<+(ި/W?Nr%(|I5#1.72ݲҢ US fn7̮7'98efwA<݊:IBd+-TH[q@ժO-18?gDG} 5h/BP4!HM "jZNH PhKP}A#9FE{1L߻@gi!N82$bBNH1"A43I'ĄsH虁>叼&_P (0~̘B)ՠig=} "L ?Qմ)'mX+KJp^Ho R:Oaf;fgkۑ@a>e}.F%{v&32sW~n5g |аm hhdsUT{x)xG3UVmN[SA-v-gq"r =e2#Gs}MY~&3,JpIL-ᒅ)D4 Jsr2A2yi!EU`sZ*s`yeW+Wáf.Z+'Q(4KL[K9DcPDL #,5挑VPw{8u`"$ƸpFH <eFianl53]a$#sib9S͚VPSϳ?MD!y)pBIjT"Yr^&5HBp pLBV>aF|_{՞3!Rx#K/,FU?2uF{7ZXLnt$v^r-~& ;Ke'Ò~4C~4>I{є4ES]Ӊ% pvaXZ+K9􉂐Ǧwo{liҷG Rվu'ggň=v՟ħjjv݇ۻYk} 5n4U!>jJ`׵vouوRM (+)YeO qF`+v 2b} Os7RM/7}hw 7A`vulmzn+/2Ì~u9[ ^;Ut8q U@MR"I;  XOXуVR$"G =ƤQxN`5c-)CnMx Ŏ M߭n&[((B1]w}Wa 7u౗1&IZV_Y]Yf4AXuZ"C"3ͭ2^@2zJ }@'4NTQnR` 8ռϚv_Q lRrJ+CynH 3SUB R!") HgW6jI3B$z6ht40#AHdKF@BL% O )h!7^7̹[e>W:$}W;뫙'aRE j6oMV@3*k\iXBﳛP+~LVWO $w}6Ibu]aS%%hӷ'7_UI:E}3q6׭ZLٯ0p6[[|UFkR Tlg__}W m_SR̦0Ka[ Z9Pz9T@HBCcro*+C00r9/& (Mk9v;QSh9\5aI`[22_0u+5`J4i;-d%j0a 1N1[ "0!jN[ 憶;)8w2Ҹ7`w1B_FP%vQgWņZE% gCߵ?GGWy( l%ZQVi㝚\!N'lׁ.SId`ƌvQp S.E%pBSL@:" ٰ0!DF @k\+qZ"R(TryPhK!e$H9d Hj [ҧ=aaog+J[;^1|䴒ur>L4c9]\xr]ǽT2>]ç|t<|Zgwn,v.qb(ԃttr2p+@ৌ͏5.]= ! A|A`7ٶyF[I~Ԇ`[RYXUf Y)Q `z4]"O O)NMҔ?Fѩ5x7\b5>U>'F~PCXSjK(4BdO!cW/Hat{_5LqR/!apX!@:4E1ҵ7Gs#zAݔaMQCkT*>>hf;do; \ ۔~4 =PMxB<0u*>=m`zZ bi6\{]>;7ES PE%-U}'ż`3!d_+U3&7ʗyw.i<|n@%FNڲsPܴ ZBB(sBiB&꜡$Fc[҂Vk=@{gZkʷ{SA&%CEg~\Y\^L6 I%_?fRi3{@d)#4pY&j(N0<7w:=ȕ1v^Վ'q3W.8tȣzLb&Vf ˹_\DUZ. ްYeeNq)ˋ"偂o7}gΣq8<q،P,/2e]x B"#. }PoR00%R~B כuaV^,ݬu짪o ɬnHYUsVAä.Vg H,vRYcKEz)FJ^WNn^(Mp:kypqv Z,ɂT?뼴ș˹Yv.fcK:(+b!)B.˸k, eNDpA'ZxAXR ޒ=YpIDΥ^J/eν'#ҍ(YO}Vx-F$G䚋Ln$*[CVݕp(J@ 1[s)Nճ]8Gn'[֚cibkAjnGƔbka憣!93H!aXlY=Q ^ZZo1iG |=V\/?}IwW#AnX< -}@PV mO=K4Fm0Z3Hp6rB)+ȗ4Z;S ȟApwZ$_:ԭrEفY~~ ˍV4aN,|Xʇ l|~{1@oJ\bgť6hyw7}A_B_t٢$e.rwIx0G?n6WuY HkJZI%9UщdjHfb,-Jtrɰb  gjvڧS%E?j93S2+-:cٞ9-Z;Lw[FU$7ަR>.0 }0uf?1KcoFCMA gn> _vtMؠ2Sj9 `\X_gЯ8Gn;H#sw>tPEs ^^k@23穳½[ҳ7+ >-RMս'&PHk /rrX!GÿZFv*#b& aMˤBc*&PGG)BC8ŕBPMAȩZOW#X)<-4)7=h@1~=dFGxӞG|P} X=2Q&Wz5eBr?ޟKBNAv rP]jAm29g Z'-A_ucڷ0sd]^(.ϣ}}zOPrr`!8Թa "8(!%s` fc=mNԩMɟ8,g;H$8vx7F[:Zu%t;%ڝ4Gif_˩v᠍=\c!ĶǠb&ViG/"T%t ȭml; & zΥ SP rK_6h*h@<4 ׊(Qksne6B0JD{ iȮ&sPxʹ9lP 9Ө]LFŝDճN=Ӣ?31;'_/v֫uiP*&LOfb v/4HaRQ4UWMO;4J+;h.Zuyԉ2R@>)tcʪgU|OMPs-m->l11q%"M|0V)A2 7B =C9e!BQք(A Y sE] ĹIxv^8 ڥqۻH%#ΉLBy1&6zR{[ ȯ:WNcEkg}si)- RX`0EB? \^[)ɻqv71sI$Ayoz[Ocu<=OIE:Nٜfonz9semz&ϋ ȑ($ZkR9V0 K D069%N,#Z0y_*d%'WaWt3T/o8w0:FCc{t̲7smQ\&GCTa42M8I8\k1q3Rb'?F5mI*xC~VVPKԩ'?>_tSx~ 744 H.^ݮoWo+1wyړ7$?pOz m|>&gPq<_mKgeV%y _}nnMTW9NW4rEݩ]]e2yW)('A1@)nKjE⼼*:,BX}o7uo\㯗v!=ݦ'w_y\u"Z JtU$Nl\P1BJQT^U~^g^}2rƊx-EgrY8JVq4 SÜ0zFk&'l *o"Fcl~iLty(Q>Y+JŬ \+.rn*@<B%+e^u:yC>x@EPsPk~.KY)鲮\|F@" PU.2# ҁsZJh]7:wDr-{˨߆fQrvrց9%8-5N}szg53X}s4ܓΫ봸Zy;>a2w]`&?Ii+wԂ`ns<:߇"$l+Nxk#zS+*Z?_3&^-Wryp/E9tLӾ*Y</ \3TqE~z]O2fxșE 71~狲c@fi^P2xǃjRsR8s!k"94L|} بUm 8-r`cpNag_RX?=8B+D9lJS(0{em$! 1'} }f`% L3{ $$d{)ד'Gا-᎚U8 CzWc^$Ir mrvM<bq 3Iƃ# R \t-n&S$f ]T`{y6qq_-|L6YLL} P$Zofpf,*w?WW Vߪ$zxZTKsn(~R'QTPK{~,V-M񇖥 |hJM.mz^|~1Y\,npY}F߲ÚBe84.YL eo]lܺS߇A6.ՍZJ= Dg+)xbu +]!4310*d?iZy ơ4{4_'3" z)׿_TJjdͨ!N+_ ƊDS-]Wq96n~ P03ZE% rF2P|ٝ)gD 5Mdld@2<{pU#I#I#I#MlC.t[W3 ,(-A1@a 5RO}v?ҁP8Ʃ8PFQ*ݼ#Ōn Ņń$V(C!)o(FaXsp:\*Z&J8}.w?{P҈J~D)*C(Eud7=T@=Bnmi"ƽ6Έ@DFNm.3s62"X&3"'n9:Oƚ%8,Z9 ȹ2k P.=>06ԜSZ7h)C5ȐH) TY뙌FA"f+0sn*-!zF|CR(6/-A GT$rrA9%`jZB=|.lm$+7n{[{qd}anNk&KpuHgVMvAwR^sÊtsC;vP+H5#H+T+2${A%Q T;!xHjTn/Hj [x.HI;&diLꊷ:nTV _c%C4Tt(13mڱ j: zL{[R I@ԨCv ゠nQ.QAmNR}}r9z*CTJkt$lv>C=XEMyyiE.!%d8]5OKu^%g=[aMFG5Z >փNvgEf3S^6z< kE_) k'8vvg;7o DVlht=:]J5N/jXS-@S.Պ4nU>\l6k^p{96:_xQϋ:q||zɾ*Wǿz}a.5ج}6)tUCX>c&*21p2mfRh+V|XkOt*,P.׍|*zNi[=spG}JNJ|%]. ي ;v`ذ0(婻$^[ѿ: ?I(HEgVz&u|ʍ ,w9-?  ^RoL\dvRIjD[ߟ:ӍLި$VaF^@_KMe<:K$O hv8T3-*<(]d%Ҩ5TxZ@WW& El1FSS(յTEq} v%<. l~} 1t{:}Fq8=x^WAPy c>߫9NhȝWǡ ٦!Z2e2k ZL9B5/Pw5QPmmVJL Kke|d3% N1 cD꞊ t"{2:ɑ#]gNG+:jl=OCNQ*tmFLLW24}8cW}4i7'@:mMm)kCQtQ;j\Xиe.߭k_f괥U-|[;dfeIfBY&.;[X>Z Jx&!3Br2D:TԠ]oŒ]~X=SEm Ede}]6zDq4ť6vrJS^tjCn'hŋPx4J4#U%Q+O+pJʕT8•ы(sNǁsz_>L>. pV!ٴ8u*KB ~o]Z7\0^B>G 9X > .Bz;rv ,(tc,{K%gp(I5}"$ Wlt%!78[ҡS`ti3S;ÉqDG~JJidRPa&u LEhA@ bVsxd"Hrʟo([̷|ʻ@RmS>C̈Y5fypB_ ֵ5a4$ލ 6Ry*Zff)fLd,Z#2pn `.>Vr\0M 76,ZK4 l"A#S!O3GX&#Vu"#*l+tn 0* Og0c(m鼴N;)7_`&‘HS{^'zuOPz/׬^ܧ_&Ƀ#y:sđ P1w%̩F/0ӝ+0Aӎi:|i5"j^"Tn`1.OPO)2&(PӸE{̬ W|e=.7ӫp׶r5Z2YW3N۹uyv6@C1w%IJw>ZJ6<*Lssn]{*Ӻ57,-U!b檪m@RA G[xL7xxz?78mR=IŇhonyaz||ejF^+]{E}co/be{$yn b>Fhx)I~XW.ߦUTނfbRjJb*-p]qj394c2.22C@Ă'Ys3 ߈6uӎ|OT;|rzB۔GjlF8A7Gqioӏ%O铭jMhy} ϷEWzHs:f6|@ ~.7Ww~K7nf%e51ߞ}4w|}Eo Ot~ýKGCh aoy3UD4D> tУ/Tc7dtTj;C}@F.boVPt?<Ք?7_q:W6s1dg}1cqKt Ox1?yjrR r6\ym-PzX"_#9|(0ŀ`>=j~) ח~ pQ`{_{KvvǏw,Ifg- /4)Rl:<S;Rl@V1)F꿟&bg !bLӛ^e2f09)O4ZoWX,CDTǿIm^uxtxtxtxf/jYqqiԄW^:thiYG8UKPxvҁ|~rsvroU˱q9ݼJ%}\ e3egrw`>چ0 "J-~YVd+CLv ^wYPb ^vkJ /:7[vkgLEH =X}yZC Ѫ<6o3e3Bp*}@Im *Q٠PA3C#3 J7~? V\FtɍQ7r9[d|rJd <<<̀5 .ZVO(%wTpqBF>֚ %DFҁ>\K'Swn&fصhO8=U\cOn\BS:'LTG~jc&Rd39ATVlB=`~{Qp!TMj)>g5;o5T3WVFkMmɻH_١ 8~[Yzh!&=ݣeH`"Ha}>K9eTV1vh|&1^[@uDZ28F">qj%'xqP"%FLqϠe?}P4ۆۆۆFVnAD&GI*1f 0 Z#'R €;(*:Q  <{6&i>1QQjD>f{ B D˻XdnfdO+(.%~A6d2ӉmԳiN"sIZu>-)~]*Ju`(ܚ%V61{}mNczWn[bj8hGǦӽb e`{o?ުf;nEJf 0s=_{`(&;`Ǽo l8ݕ=aGZ^A}pԻ> OE7'_;x -=pt}Kq]8PK[!};=~xx4$]ƅ~eҔLWoW{ jnwӺ޸1wcm_`&-@ ۠?K3@28T']86֥}~Q<>\.2(e9zwuY7̗n5^?~Wo.a/^_5,`^c~qvjAohy8}{8>Ab[3?ϗ'orqqygXNylW:pV˽l:>wrz}zqIqks [0tkOagl[Y9ٷɾ.okI[ća7Ö:_] C^8./]м&3R,4| [al10E~={SrO+GgDQ+vmzͶ-->6w*EϺPBHYa\08hϠCQ [^<Ϛ3>)VQ=PFAҧ:NgPqcL!5ٿL*GVT~RzʙM3ׁ~cَ;0tὤ9Q{ԻϾ\ ##av_}y@鷟F26mevrPnV@Ϳ#ds]0z!2RqtǞg>rc|r2(&bX+KLN(GJ\'y8"{&(<nvm&䘏k|w1y嘥|3'0}צ&X 򼃞 >>a ~J\4뒃3j|wo/ %x?a ?u:+m+]6/ϗLo0RC%8+z|]SNzœ] s&aw%O"Cʧ,);mJ8C=q0 n:, C d|ް {ްi{ aEBcDM0 | * QĻ D@N4s.ݣxM`R2i?t.;<+h 7*FP@CH-%0, 0"Ds% *+6PdsJx-c"4TW'.]/(K'@#%y3&'"1DCtXEHR-xbYC<=\E'i!Dϓ:Jc${*;Wǵv, J/y[حϡdG8yòl N;oػ>sȬpZ.Kދiha?w:; *u. ( uR?OEq4Os؎{ykc. Zɶ _45iTƖSk&v? l6B?x=VeUa >9p*=j8x3݆,RK i&saU՘SE1&W=昲ZM0⪽vU.@*H"efȀ3!xLcX1*Q1Tɜ,P T{aW;Hx{bǐ Úgc}׫R&jXD3Z*w\ҐPBt܄]fNe`j[cܚj?DS1u QsbR_s GH!E97W 7\Sf/_)KϜ)m 3<7"Udkճ 32]F?F{㳛q~)%"~^i}98x6S*ʤxW`\1X{_x%=s۰Љpb̀_($׬(So!~CwA DG @ ( !눳y뗇 Lʜe("iO-:#kė*!4%XGjZ"k6 8G1AM/E+TUٗbSr`R$ncUd`1cwY*SCE*AH=EBQ#d9e:N$|*RD24HLcrEЬġBAYmwx5"4{cyo 5'Z͈Zy+V(v.(APkEr(V>%1!3=b-VE}R?o(!k4;{[!mgs' +4(]݋5+a 竹LyR]U1#<{0.p ;z0Ko~_oQ |y%i%i3$3883T^wNI 8L/k=ӕ1J?{6y~vսNx`/M3v{E%>n[e(한[ֹoqtBr9&򙭿9mlERAc-BYؤ[seQtx={^Za+_\Ppal (*L;,#gN R<\  \P hXvwNfGnZoc.mL.{!ZfxFr [ӰևZM|c5+l\xrY{ë撊•5kC0Lß.AMKdlk 3ڑm 3xm5δt؊M]i`; ΞEH!q#pgc,O9|@ֈOj%Sal!XpEn d$ hC\-vS?V>St^Rlք>eZT%)У}{q^m4vÏѬ| n F4fb@cNmJĄHʈ@HНG[ ,C'p5n(b=%q)E)lg%f _[S4ô9zê|$yZVP E/ uC DDBJ 3&& 0: 98*$^+pՂaT€.t@l(!H4@ -6 JG,:4izRC^}AĊEfNN;]tS {@~4{mh j5ՆQw"FJdOdD,1_l`W ּqknwO_57[D(Q`ٰd\)3e$%2_PU)d-2LH lӻrr+S86J _iPGkMpMқE&c 1V;ſibOwnw$ygc놣|! .ZO=y vstFըWq1('2@kJWӁ|h`DdmƗV1dyuZAǝ.?_Q( a0$(0PTV1)I{r4?id 5zmX-eۼ;#iG+*K }wyӎ7% I ]Ւysk=}?0 po/Ƌ|6|6|6|6" bEcC4ѡXc G"c($Llcr!HDXzxQtB7^|N>@bl=zT,JME#:*T!^Ԙݖ4*2H(6(Z*h"8M8Q aaJ@D8 ) Q3}ar$"b00|;bֳ. [n5[0)׻a}"znHUQտ˾R0Jݐ+Ap[],ysG M{ t%x<Vl4rtKҟsO cW936^&ת6TTwyz RW&`QeBIJm{iF(AsF^~cZE# ,6NM [uRLDaˠ7-4guBtyS@5) F E qv /š[BE:}zIuݗJc Fh"&"NAdطc9V[%{kbYX??%/\՚Tg XaN2 c[ }CIшWbB l֔u3E֜ Sӭxiɶa[A׏3#dvi sNo|I.e \?֋+x,=l`)5ݲtL ̎92[t_WORM}3;g۬bK;g ~1*v0wnLğά| 3W>mnmTvN ˭ O$b<ٹkCɒٱFON-(;wj[^PQ""k',9۾m%Bp,){iE[׬Zb*y'aD:@Fxс<*ɜB`}* ℱRgQVhq"4AIeqaRDǔ/HJCW/˼.*ݳ\齚# %zRǾйf)~2b]u9u 4ZxW 6,XN.`; WrZ4ΦkIR]Ƚ}w5je7su _5Ӹ%=mpLs]LӞ>kULzuk͝n)1qK7 {,Ei sQbdL9n%1QJ0dIWg0>&zDžj#P`Ābs,,P(p[#Is餺ir3t;iOśloZR 0acDBR"!: s&Qd_=jbrX4hR[8Z(x$.ÙɟIOk[=Z$b[h1d7DXh 6:S׳E4= ́=֧c 2}z%S2j-p@aokCS%{t/NLZW ')ʺ3]]=[3l.xWr-#%yWH )6ѴC>i4$J ua+!ӼUOg\:CZ9Jw/Tp (IEھd=渜kUsY]ѩeHOmߍVn=^.31[WpgI[ O]ZR'UR:u?U/9 V|Uɨwf7Teo>LfIk᎙\a E!(F#oӏG^k#!/v.(Hahe lnKwweΰY{Ħm9ԯːŴgNR5!Zq`BmD)mCn`xpO'v^3 ];?Mѻ뻳9?og#o\Nf_+za/axI'idTo}]ͫ{%a.UoOaryf-w pplJ;G,1 68!A[,U'}m jJ4t/62(yby_g7615U[FNBShv+>f5Cr3vfʪpVm~|@l?\:Y3X?)4y_h7~f"e#ҹ3Гѥξ\M5X6 38y2x'^]aSiZcBA9h~*@3%2'v/)d&KIKPri`O]*sjL(>fa+mȠrLBijѥuG/a`R`)Rg)TRJSY)U$"פ5nvHwWQǏw.,+}a ZH=Sv?>%fQvwX~S]b *` +\I *"9}f1wpxq~i NIc2׬v~).h)͡18q^sX1`@MXTKt^F"\s'E xcaZs3C8RfHnES3`x=j4J*HOo?x\ 8by;MO34 )<@ ~+=|k!zہ*j'sKK3@ 2P~mĜ+t!@4vk0󜂳 Kj\B (+8*2^bS xjNoN:m+Fk;AVaYn1td5lHevx*q\An[=1"2ЦCb퓣ĂlDWMWG `j6ٗ-+-}7H@}{r/8(hգ .%z% bΎ@)4b6BR->*~SxfOQfzZbPA!e"BGEq>p2E:@f tcpG.d]_H o=dc!v:[KA͆ 3$"W *VjYt IVT2LAULT%j5NF:䄯1ZkQ6ԧ&-*tj 0~0*(*]Wy^ L5_&׸~_]DP<B7.D&\F2^q(52L;ĀqD"2I'z>m\DB,s'#*YT`2HYzRY˸.&5 @Z}?!MtAPTI|V`Κ' .*AlcВÁͮV38'% &Y.+pRT0֛)4m |H*njS3Mw4(O\B-ut/2};&P9Hgs1)җ NɨRU; <ġ[[\]ƭ(ۅ]9Ov<ٕdWv3XtCQвY1R2FU2Éc.(ϩF1MKeՆ  »XbJVFzël H-AuwL'wN:@fKIV67z6ma27J/X`}.>.:-/UF.?q_>^4*ݽM_ HӍΊju z+xx`JD?;zBIDOPG gNy#l 붪fRTxۅmY2w/6j%51k#^6p0dQgݖUf2l,n{կkhPmh*S(o%уCtPIa"^TJs#ЋZl Zp^J!jYZ$5t=63C’Wp= J ir@N7p6g͎b:N6I( ,wY.>svϟbA&Dt,~ƶ2F D"om _5T2Mfd.9B ̥#\Ym3iy¯Y!5h]C^tne`j`7S})?+>^4Ȃ/"ۻg./V)a_Ƀ4XdlLH~@v cDɞwg8m `ggk2=ݠ mm=j6Zp1h[4ݏAX\W G.l#mzs#Gm{Qh6[ ѓжV~ o6U(nҐNh!Pob`aFw\z;!= ) h· 7%ќ?_ r+97¬śMPJ1n:g,L.-#\Akn5gXE{X+aQs1ũv4 -E,Bi_6&f0;5[s[S5f65 o+caDzGv,&M.~0^8eE! jREVw[LZk6-gLMlZ*%^故tQKmc恚1 bjYnlԁ߃ۼy0/U).`>zṕU 3PfhplS.!&hci1Q?k4uNۇme9 ~sZv-Ɯj ᪳b6|-yo9^GXdv1;JgZi>%W4빎uWs{Fw7D{#sIY23+" rR-=+2AҕڂE-j2u0LEuӭj ֱ#=,6R\]s7WTt]@@emʮ$ORk->J$_zx(- g83;~V &%&(0 (c!eH5,K:skPԈg lQK@H@J$ Igt4Sx#%rћϧez;J/%QٻOSY^ޏR.$%hU7? \͛| eVcZg?]5w_6r$fdWwxws~Nƶ;j 5lZa]].}8A Wƨ̠epTOw4*jHH|cO5LhWh+3в mt c')8kDЎnLgYv4uVkݘ9/mm%ruͣ2yKT!WpuL:k5́ ,L{.f \ Yô^Ġ-57 @kؠD7'm`.mOhSנmP0Ȥ8]^# R@`zUS DJS(@`Ҁ6 *``gB[TKi&w kݯDs3mGNфr4M8mMho#Bd A# ف&*&-]L";7..E}ΒDmU\/d$c{ﴸ};-nikJ:8n,:dR$O#cV'lM`0 d,TKz/$RBXp$Fgx:FJg1MB8L' Nޑ4JvRz6[t!}%SBx%9sXU^Dg)LJdE@G.)FJdFIz X?K}} J"-u(zvWKɳɝ_W ԃLn3yeM,flBsG(ee94Kn67]ӳKlA,55Ř2J߯hy#cb=,mh.;TVaXlUz籧N~]J' dZʷߗ$>h9AQh΃<\#ݟ,qNf=`Ǟ2E3Ww4:i5dMcLlN2q?rW2Bs)SA]XSIWǏ/:ܦ"wLr݅;{fS[-eTaB7q$ )kv5Uc5s)F^@Ƭa+pp$J؞8@2hLH* X58m]cJUfmo)`ߓhpsrן~l/0XdM|[\۷9[.7U9[?{}勓w7]~eݳ܎g3ͳ\(3&X&o"qOi|T|cZ(NUY5Se|khҍCLC6(L 䑾>;{d+WO b}O'b:¾sGQL˽Sl 7JC$FAg9mWZ<ƹ}@4^ g=kGCw'9&{zm(sUB< < 2Qzp <.#3;{[I]Q=J;upT\~-081V]VĹ]>SEAyd F+u JG$z{&h g*&NmY#w1!sMTI.TR htpd!=&eNX!Bbmr"싸NX .K>uf[_*s|Iޢx+ Cra7?)+#`j5īeZ}E9_|9;zҞ+;yI>[^]ͪ{cwY'+o\ gL/$Z$ubvr~ףdx+_Nܣ|<E  ѧ;6xsyIcWt[5-AofmVzO޶A%L "`?dp"9S` )1N9bP Jz 1Ls"pY !8KFfY/$yBĒE LMA`mk4;Yp pTIE jqt)-wV6ӊO|5hYhA.K/`+~!cˠ (vCdyjyD3]IY kn1ӺͷL+`])!̴<A1mr 9o#tKf-+U.:!O|.*^orPJpZ-tL<* %Z0q$$-,/AX\.piENZmVv{ jhC.hO.-5u1N iC\+P`&B.5$X0AXB06i_xg'ƠHkנR?!tj[S %0N tX+\DRS:wK@FbI0E.bF*A@"mTq%l३KTi# |z{j%Γ$I$aE*-ɞޢ`erI{E2]3=x3Do|9/IzX/2˳kvv]}(7os.}w)ήNie ۡVJƜnhw_ 'W*jW x APl:_G7ƻP~w? ׂ5j3NaxfI>W.ཷ?{pBofGw?`LÁ:kA~Wc¹Ż Wvgk?щv.a__ujX!:j"pS5nZn7.o]Ck;[4F:hhՂGAb" ~Юq֮q֮q֮q]gWJ2Ayׁ+c c%(M 3c!P !$2GtunS_ہ>~Z| Gdq7˷r4T-c6rw*ժ\,QuJlKI̥bLb .i#S$-fYkN"չᆪCE`ThNK"!k89aro b=L3`jx9zh)Qbh& DYW5U_"PYf–P0T(,r',2ƈD4`(e)7$Jes.O ˳5v|zD!"SM^#qB/4QV;Ҹ$s?OF@m=Yx4R#a+הQdh\…%] 2CC.8-pSgk|+k!Ђ2T႙㣡 3ǡWUYvъ ^K?NWY흛mzrdºAlL[ʝƉ**ƿr@Je0T[o8/Vl!@c.t!Ta.mY߾ՎlR/w_E= gٳC !Wcv >`\6Ў1zHX4׳'NgjL %{Fe; 0֕Bs\z@tC#I:պA3d_8QPVαJ6frϝ?Q;xR4@˯g"zC'V\t7f2 #Aj)AMGRN*h\h ae+2Yh-NbŃFFFFUXiH'!hF"xJ!8M FKmJ F{+ԪbE_ZpKݸ*_`KM&Ғz;C.ÒDiD)t*I] ZT:KYr9 jd">&o $%CY(CNj=DoN[ϝ.|%@?2xiE`h1A$FV)g"=O@wk&&)O^[-yPAߎ"^yh+~Rbknl%*$b>~~8hql\蒇VΉZO*΀/7o (n;ïyf^*4A&D;BqE|$\Cw]0);/E;QXoIц* D1sNa#H 8%c Y "r}>HsCUʥ̎,z^І#YFH Qw&?bn\I}ؚ Ex-g"zU&o~ǿъsIkTc-k֑);$%GfAvx d-m ?Vnl:]TlъX!d>}`; (A6kj1UU>]Tc BL> &Z)Awf7,,::He!53C4k~pSÙ O1zF#ZWaåВLI@`M@:~qqqq]?g9-$5R 5CF͍UʰE܇\Yc|HAAH<Ɗ @ΠPY| T)K1[oΚd`$Nh˵cNInȐ'ӍºK~Y-vB!j>ҡDFU]zBo94~6Z9U<ם5rT:S M._Tv#N\ja~ouc)]/`zs]煪LJHʻI(2 4 ^AiBqm hF0u"&8 N<%F4;%NJ Bƕm-w-g=å !* GV-D.å ^)SL} Bå)+2:IAh2y A'KdZ G5~ 8;̈́X;l#95wm3rZΊ0B컣܏|T_asTN0]=}Zݟ=\z"&VǴoDXn>wp7 xGtgJo>6;8~?.:5v1 ⮇.*";sǣr R;//snSs T .LarHQwwȻY_3h&[ H1rbuD:zb:ޡDѪ; :A1Lu+UT{ӆv2-+T}䠷 s-J(̃djrS~z%=zÒ3f1=$TדO%-`=E\"ߧVRkU]܌9<)+s"uNo*2$#]j˺B{aOw.Z-kv5xbL 4Q8EP02QƜ , #7 qWy/~#u8 "(%p\|c *Z*.%M(1$Ng7!еK/og/K vz;QsTb"A'tb:_)Iy% D o<1|49ʯWA+10DhlTV$D 0υԶ߀.9Y. ON*$sq; S $M:b*d9h;?SrA E\WD|UIBjna,gSd*p2H{$Y5z"֒ac=1\=[WhRK%ZW0Brn op>$ Mr\ ϣ*ֹL90q( QdP+xZD4 R{bbԬ*<@Ar1)r`>8>[5]Z%[;Zwg>3[,+$~A\oqcspG{+t`oR#Ʌ<*ʰ`FIwn_capmm];[&QjZQ͹GV[3 7[Kӻ;J*=3t:Hwjv/}*:k)b3no]G]ܳ79[w@dz>ioZ+QooFmH,Z0pL&{v͢\rl}w>oe\fI2'QU]\aEZca2AAU9 {JrF/ȫS~;mJ-n0WN]%Du ;v[vwt~a7Ow/} Zs&aP;Nwܬ| Av^޷ 2zcbqW=LG|AZw}qHO6ëjD0Qod0]pN\b }djʤ3\ -Nאr\(b|t'z?NpުBcɔ ):1}0-k_p ^`kg"z}lG!: lZ zc;k_w_vG*\kׂ<ݞ ,"C}vUɍ]́{4έy(BMݷ>olNwŭ[qS=c!@/ӑmnŠDXw?5[rzM'"ȝx{z&ukޤzwzݍY6GizξIe&5 BfQ ^Y'K7E8Ty:! _vw1;{g`DBS'wvvm;pm]5*'a6(eYU5}}o{^Y=NsPZy #r˳x&_H)&?xjVŒte5EbmL;Z*V=}[O./ND⹏|%+ϗ 6ph(%rN< ڞ DW.?-)tdE5o4CJX"㌚{7Dm]>ģ㣫'{ yK&N8%)fJPSX Az|@14H}1k_:_ % \H_Tr9G%J-;?IaQ(P:쌭፺E\N״Ymgd3g+R}lINEg>7jvY GѮk)!Kqz|uZ8]qlLʮR2vgݭ`Sgݯڲ.F>r5J}cGycdgC&:M p[6Vt!ӇE(%+҉A$tIIo\9LvA`93E7]ȫnS!8ġ6kO5٨+9{Z7ެc=4+ޠ riУ1dTv4I2a@xp7_ȪҨu$e _ui-{qfO@F2YI‘zi($8 eB0{4:L&N|!6$k B2HN *ǔj_ ia'_q'<'j.@۔FW_A~k.B[sd,1A\E:3[\_`0m/~٣7v¢N6\&SP*2ƜiEc3 h9GmL*ؔ޵֑#"e`,oz_/  мy4+[d=9GRD6}.3ZnuW}U,YŏH5,&~0ᔗ3 ޼/:.>jH?/;j8EL׽g3cC]VY^uDmEw[r&Ljv=1Brf2+`068Ti҅JYPK,u D{VV%Y{l)RP 1KczmlU.?\ٙc7nr1د{|grJtc1ӗu1{N츼mn9pJFy+[zL4Gg2m 7u)-Gb58G &['W)v;w;RY ZƷZR+B'$ u?@jQR BvaQٺt~M')?﫯Yۥ[U.OVcmmǿ9jq_{^G1|=8l}vj+KCn=ԛ?ϫ+ru>p)'w'j1qoDcγ hnuC^ndVP!CqBEȬRBb1@`**d=J*5UϴŒ%_E s1˿:Q+c7bT_Jsyg'yoM T$} <(Ar> <|DHc2MO 9k6虮Rzw 8i`N *z 7q;q׺4Sr#1mi0EչQ'n9(96zyeiI2j +8j<-@ >OǸf#j5DAgk4.H!@ׯjuk87|xO51^dGq oŔ7 ցkGNtq[}ACuwM^\@p>sٸnx{E~cv\L|&q}f1Pd&רVj5 ok6'kʴA X1i+shk*"81q\]dA\vBt_P'/my8 BtV??Iwl"||1_t񏷮y\1 u#ޠIv,4L{ځfHJ3j2 5w#e$yFo'TD4gz (|9VfH+`ԄqSޏ `bF*ɏJf#,D#gQDL5$.bR2FYK$";7(Z28ϼa5 0& !0Ήz[X[5qI G9;q\l6~7Yk Oc_:d/W8ApaG;)RwwgG"ׇݾy rx|ϛBAt9E5.w, 4(F#~I2JPő lX%/5K0%CBAN7!N2䏇.t!`sq,&$3YoB:.q)2L[">?D̜/b*` f 1J#G{rp'5Ƿh#a8#Yg$Z8 odY0 C.(3| Sk̼jM!O7,VLpF.MYzzv+ emE%d|Жnm!)X2ٓR@4!Հ̒M %gMX !tHXWS+\.V; !3pNvޭT Z`z\)dzaF.K]!#dsnrJb|C+1r ZWItl 4j$GW^+,Q@"BcB1Ǣ{yqjrd G\^ 0eq j6(Tq*@ӰAҠdᜎ(,wTxO>|ؖ:v<d|V~xCxhs\]Ƨwu[-9Sv+lѭZC^V`'qD7$#FZ r\%혫þ[ѭpS( h|@Cwa:Ƹ<8tl{8C~ؖT[rލj_H}_jkw3$S6}EQ4(c5KG7vgQ8c H;#k#鐃|zFhbDtߖ@% ~j(zCnY]\zð'X 9A2^('HkuRO y-99o9 `4J yrx/݋M f|:򭢔>B'@RQ(ZBT?n  86#ϓG'! A\8oIm K,qqQylaR zFHC6`퇪Hk?,ӕYj`J?WF&d°tw&>1Q!f=<7^B΂#g *;|. 8piuOѼE,ܣP~pv`Ƀ0m>sǞזCZc5̡bI=3quN]qB1'HC21x> g ?<dM0 OwP ƄpSKpR\rG%B&UB:c9 `Ln,ǐs Bu ]v@)q)<D|S .M*Xgk ]0F$`0]g!ǝEXX3K#0vއ5\ =c 6rͷe.ô 3`0kEY&1#`9 JԂ𿎘k >LC^?+3C fZ\FT`$< gN+\rɔw竡\Ny03BlqGq"tPD=.ıżqrb25XóX b%#MVOnb4"/yCXUW!:uNӓq֥kԍ 4/xԍckC524]q2!zuݦ @غ_,XmR ^$uuͶFjF-V`V38B-5T{p l.Pe~x4t35K}C^VX&; &Z r\%xyW)xF!3[ y.ZS3>VA~$s!ĀdVh#3[>% 6L|y{1a#P~܂ P6bWwp݁3tϿon}<"i5iJ֤a}=qͬ =eʍ2hr(5J1ˀfCXnyT4PҁIMC9T(rKo)$꧳K(֗ohԪUYLj Ev^W1~L o]`6|rKh/n֣,J(1-tM|e/1Xf42-z8l[/~fyX|џ?7*gY7^%;"woIέTBڎAzO;#4P4= ;[K>6L󹽦zՆ9K_G[і̦2k#!~%9kFJhњu29idMLص3͇kB&qOLZKgXɐ0'c*vVBhOq6֟XbyX?d\VHՇiqT@{ Am!>c=1{%Kb46 H's\on:UDTR=ݼ;NEj #PdH9qs?܏QX]r6f&)WWEi ybxu PzxO(yzp6RV |I,(/"9a*e r1iP]rCΌӛNUQ"َay;AR dZv#Fժ1 zm癇e.;m?L{¥<}>var/home/core/zuul-output/logs/kubelet.log0000644000000000000000001751367115155647460017725 0ustar rootrootMar 16 00:06:10 crc systemd[1]: Starting Kubernetes Kubelet... Mar 16 00:06:10 crc restorecon[4701]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:10 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 16 00:06:11 crc restorecon[4701]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 16 00:06:12 crc kubenswrapper[4751]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 16 00:06:12 crc kubenswrapper[4751]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 16 00:06:12 crc kubenswrapper[4751]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 16 00:06:12 crc kubenswrapper[4751]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 16 00:06:12 crc kubenswrapper[4751]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 16 00:06:12 crc kubenswrapper[4751]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.164131 4751 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171802 4751 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171833 4751 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171843 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171852 4751 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171860 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171870 4751 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171879 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171887 4751 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171895 4751 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171903 4751 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171910 4751 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171919 4751 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171926 4751 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171934 4751 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171941 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171949 4751 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171968 4751 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171976 4751 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171984 4751 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171991 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.171999 4751 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172007 4751 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172016 4751 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172023 4751 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172034 4751 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172045 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172053 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172062 4751 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172070 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172078 4751 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172089 4751 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172099 4751 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172144 4751 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172153 4751 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172162 4751 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172172 4751 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172180 4751 feature_gate.go:330] unrecognized feature gate: Example Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172189 4751 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172197 4751 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172205 4751 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172213 4751 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172220 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172228 4751 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172235 4751 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172243 4751 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172250 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172259 4751 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172267 4751 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172275 4751 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172285 4751 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172294 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172302 4751 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172310 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172318 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172326 4751 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172334 4751 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172346 4751 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172354 4751 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172363 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172371 4751 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172379 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172387 4751 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172395 4751 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172402 4751 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172410 4751 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172417 4751 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172425 4751 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172433 4751 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172440 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172448 4751 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.172456 4751 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172613 4751 flags.go:64] FLAG: --address="0.0.0.0" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172628 4751 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172642 4751 flags.go:64] FLAG: --anonymous-auth="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172653 4751 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172693 4751 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172702 4751 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172714 4751 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172725 4751 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172735 4751 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172744 4751 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172753 4751 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172764 4751 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172774 4751 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172783 4751 flags.go:64] FLAG: --cgroup-root="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172792 4751 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172801 4751 flags.go:64] FLAG: --client-ca-file="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172810 4751 flags.go:64] FLAG: --cloud-config="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172818 4751 flags.go:64] FLAG: --cloud-provider="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172827 4751 flags.go:64] FLAG: --cluster-dns="[]" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172838 4751 flags.go:64] FLAG: --cluster-domain="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172847 4751 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172856 4751 flags.go:64] FLAG: --config-dir="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172865 4751 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172874 4751 flags.go:64] FLAG: --container-log-max-files="5" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172886 4751 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172895 4751 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172904 4751 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172913 4751 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172922 4751 flags.go:64] FLAG: --contention-profiling="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172931 4751 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172940 4751 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172950 4751 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172959 4751 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172970 4751 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172980 4751 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172989 4751 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.172998 4751 flags.go:64] FLAG: --enable-load-reader="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173006 4751 flags.go:64] FLAG: --enable-server="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173015 4751 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173027 4751 flags.go:64] FLAG: --event-burst="100" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173036 4751 flags.go:64] FLAG: --event-qps="50" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173045 4751 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173054 4751 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173063 4751 flags.go:64] FLAG: --eviction-hard="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173074 4751 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173082 4751 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173091 4751 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173101 4751 flags.go:64] FLAG: --eviction-soft="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173143 4751 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173153 4751 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173162 4751 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173171 4751 flags.go:64] FLAG: --experimental-mounter-path="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173180 4751 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173189 4751 flags.go:64] FLAG: --fail-swap-on="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173198 4751 flags.go:64] FLAG: --feature-gates="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173209 4751 flags.go:64] FLAG: --file-check-frequency="20s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173218 4751 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173227 4751 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173236 4751 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173245 4751 flags.go:64] FLAG: --healthz-port="10248" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173255 4751 flags.go:64] FLAG: --help="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173264 4751 flags.go:64] FLAG: --hostname-override="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173273 4751 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173282 4751 flags.go:64] FLAG: --http-check-frequency="20s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173291 4751 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173301 4751 flags.go:64] FLAG: --image-credential-provider-config="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173309 4751 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173318 4751 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173327 4751 flags.go:64] FLAG: --image-service-endpoint="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173336 4751 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173345 4751 flags.go:64] FLAG: --kube-api-burst="100" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173354 4751 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173363 4751 flags.go:64] FLAG: --kube-api-qps="50" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173372 4751 flags.go:64] FLAG: --kube-reserved="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173381 4751 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173389 4751 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173399 4751 flags.go:64] FLAG: --kubelet-cgroups="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173408 4751 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173416 4751 flags.go:64] FLAG: --lock-file="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173425 4751 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173433 4751 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173443 4751 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173456 4751 flags.go:64] FLAG: --log-json-split-stream="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173466 4751 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173476 4751 flags.go:64] FLAG: --log-text-split-stream="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173484 4751 flags.go:64] FLAG: --logging-format="text" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173493 4751 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173503 4751 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173512 4751 flags.go:64] FLAG: --manifest-url="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173521 4751 flags.go:64] FLAG: --manifest-url-header="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173542 4751 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173551 4751 flags.go:64] FLAG: --max-open-files="1000000" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173562 4751 flags.go:64] FLAG: --max-pods="110" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173571 4751 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173582 4751 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173593 4751 flags.go:64] FLAG: --memory-manager-policy="None" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173602 4751 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173611 4751 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173620 4751 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173629 4751 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173646 4751 flags.go:64] FLAG: --node-status-max-images="50" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173655 4751 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173664 4751 flags.go:64] FLAG: --oom-score-adj="-999" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173673 4751 flags.go:64] FLAG: --pod-cidr="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173682 4751 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173693 4751 flags.go:64] FLAG: --pod-manifest-path="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173702 4751 flags.go:64] FLAG: --pod-max-pids="-1" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173711 4751 flags.go:64] FLAG: --pods-per-core="0" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173720 4751 flags.go:64] FLAG: --port="10250" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173729 4751 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173738 4751 flags.go:64] FLAG: --provider-id="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173747 4751 flags.go:64] FLAG: --qos-reserved="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173756 4751 flags.go:64] FLAG: --read-only-port="10255" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173764 4751 flags.go:64] FLAG: --register-node="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173773 4751 flags.go:64] FLAG: --register-schedulable="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173782 4751 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173796 4751 flags.go:64] FLAG: --registry-burst="10" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173805 4751 flags.go:64] FLAG: --registry-qps="5" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173813 4751 flags.go:64] FLAG: --reserved-cpus="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173824 4751 flags.go:64] FLAG: --reserved-memory="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173835 4751 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173845 4751 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173854 4751 flags.go:64] FLAG: --rotate-certificates="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173863 4751 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173872 4751 flags.go:64] FLAG: --runonce="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173880 4751 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173889 4751 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173899 4751 flags.go:64] FLAG: --seccomp-default="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173908 4751 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173917 4751 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173926 4751 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173935 4751 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173945 4751 flags.go:64] FLAG: --storage-driver-password="root" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173953 4751 flags.go:64] FLAG: --storage-driver-secure="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173962 4751 flags.go:64] FLAG: --storage-driver-table="stats" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173971 4751 flags.go:64] FLAG: --storage-driver-user="root" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173980 4751 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173989 4751 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.173998 4751 flags.go:64] FLAG: --system-cgroups="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174007 4751 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174020 4751 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174029 4751 flags.go:64] FLAG: --tls-cert-file="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174038 4751 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174049 4751 flags.go:64] FLAG: --tls-min-version="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174058 4751 flags.go:64] FLAG: --tls-private-key-file="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174067 4751 flags.go:64] FLAG: --topology-manager-policy="none" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174077 4751 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174086 4751 flags.go:64] FLAG: --topology-manager-scope="container" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174095 4751 flags.go:64] FLAG: --v="2" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174133 4751 flags.go:64] FLAG: --version="false" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174144 4751 flags.go:64] FLAG: --vmodule="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174154 4751 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.174164 4751 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174391 4751 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174402 4751 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174412 4751 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174420 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174428 4751 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174439 4751 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174451 4751 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174470 4751 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174479 4751 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174487 4751 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174495 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174503 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174511 4751 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174519 4751 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174527 4751 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174535 4751 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174543 4751 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174551 4751 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174559 4751 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174567 4751 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174575 4751 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174584 4751 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174592 4751 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174599 4751 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174607 4751 feature_gate.go:330] unrecognized feature gate: Example Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174615 4751 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174623 4751 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174630 4751 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174638 4751 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174646 4751 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174654 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174661 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174671 4751 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174679 4751 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174689 4751 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174699 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174708 4751 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174717 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174727 4751 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174736 4751 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174744 4751 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174752 4751 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174761 4751 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174769 4751 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174780 4751 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174790 4751 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174798 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174807 4751 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174815 4751 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174823 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174831 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174839 4751 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174847 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174855 4751 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174862 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174870 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174878 4751 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174885 4751 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174895 4751 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174904 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174912 4751 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174920 4751 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174928 4751 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174935 4751 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174943 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174950 4751 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174958 4751 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174965 4751 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174973 4751 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174980 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.174989 4751 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.175002 4751 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.189279 4751 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.189323 4751 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189475 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189493 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189504 4751 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189516 4751 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189526 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189536 4751 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189546 4751 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189556 4751 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189566 4751 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189601 4751 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189615 4751 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189629 4751 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189642 4751 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189652 4751 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189664 4751 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189674 4751 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189684 4751 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189697 4751 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189710 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189723 4751 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189733 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189744 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189755 4751 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189766 4751 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189777 4751 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189788 4751 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189799 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189810 4751 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189822 4751 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189835 4751 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189846 4751 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189857 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189868 4751 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189880 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189892 4751 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189903 4751 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189919 4751 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189932 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189944 4751 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189954 4751 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189965 4751 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189978 4751 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.189990 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190001 4751 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190012 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190022 4751 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190032 4751 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190043 4751 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190053 4751 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190063 4751 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190074 4751 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190084 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190094 4751 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190139 4751 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190151 4751 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190162 4751 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190172 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190182 4751 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190192 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190202 4751 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190211 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190221 4751 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190231 4751 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190242 4751 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190251 4751 feature_gate.go:330] unrecognized feature gate: Example Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190266 4751 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190277 4751 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190287 4751 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190297 4751 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190306 4751 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190320 4751 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.190337 4751 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190642 4751 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190659 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190670 4751 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190681 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190691 4751 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190701 4751 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190711 4751 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190721 4751 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190731 4751 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190741 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190751 4751 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190762 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190772 4751 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190782 4751 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190792 4751 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190802 4751 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190812 4751 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190822 4751 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190836 4751 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190849 4751 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190861 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190872 4751 feature_gate.go:330] unrecognized feature gate: Example Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190883 4751 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190895 4751 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190906 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190916 4751 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190926 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190936 4751 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190946 4751 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190958 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190968 4751 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190979 4751 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190988 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.190999 4751 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191009 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191020 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191030 4751 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191040 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191050 4751 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191060 4751 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191071 4751 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191084 4751 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191094 4751 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191136 4751 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191147 4751 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191157 4751 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191167 4751 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191176 4751 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191184 4751 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191194 4751 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191203 4751 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191212 4751 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191221 4751 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191231 4751 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191243 4751 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191256 4751 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191266 4751 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191276 4751 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191286 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191295 4751 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191304 4751 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191314 4751 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191333 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191345 4751 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191356 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191372 4751 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191382 4751 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191392 4751 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191401 4751 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191411 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.191420 4751 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.191434 4751 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.191691 4751 server.go:940] "Client rotation is on, will bootstrap in background" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.199468 4751 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.205172 4751 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.205368 4751 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.207815 4751 server.go:997] "Starting client certificate rotation" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.207871 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.208065 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.233100 4751 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.235846 4751 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.237376 4751 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.256769 4751 log.go:25] "Validated CRI v1 runtime API" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.298725 4751 log.go:25] "Validated CRI v1 image API" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.301869 4751 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.308062 4751 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-16-00-00-39-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.308149 4751 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.338774 4751 manager.go:217] Machine: {Timestamp:2026-03-16 00:06:12.335580038 +0000 UTC m=+0.604912408 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654116352 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:7dd49885-293a-444a-a9e7-56b3b0908e03 BootID:f81d5235-5710-4ce7-825a-012028fcd5a2 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827056128 Type:vfs Inodes:4108168 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108168 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:08:81:da Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:08:81:da Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:75:af:f5 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:d5:0b:09 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:2a:c4:af Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:63:7b:70 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:42:3a:4c:bd:20:ff Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3e:67:71:ab:e9:7e Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654116352 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.339471 4751 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.339837 4751 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.340514 4751 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.340852 4751 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.340906 4751 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.341667 4751 topology_manager.go:138] "Creating topology manager with none policy" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.341691 4751 container_manager_linux.go:303] "Creating device plugin manager" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.342474 4751 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.342540 4751 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.342892 4751 state_mem.go:36] "Initialized new in-memory state store" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.343077 4751 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.347249 4751 kubelet.go:418] "Attempting to sync node with API server" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.347292 4751 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.347339 4751 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.347362 4751 kubelet.go:324] "Adding apiserver pod source" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.347381 4751 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.352834 4751 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.353788 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.353767 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.353933 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.353962 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.354460 4751 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.357310 4751 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360276 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360346 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360362 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360387 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360410 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360428 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360442 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360465 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360484 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360498 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360518 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.360531 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.361732 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.362442 4751 server.go:1280] "Started kubelet" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.362708 4751 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.362801 4751 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.364294 4751 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.364425 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:12 crc systemd[1]: Started Kubernetes Kubelet. Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.368032 4751 server.go:460] "Adding debug handlers to kubelet server" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.369660 4751 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.369864 4751 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.371098 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.371552 4751 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.378596 4751 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.380902 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="200ms" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.380098 4751 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.382292 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.382401 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.382912 4751 factory.go:55] Registering systemd factory Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.382939 4751 factory.go:221] Registration of the systemd container factory successfully Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.383456 4751 factory.go:153] Registering CRI-O factory Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.383488 4751 factory.go:221] Registration of the crio container factory successfully Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.383597 4751 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.383632 4751 factory.go:103] Registering Raw factory Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.383658 4751 manager.go:1196] Started watching for new ooms in manager Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.385067 4751 manager.go:319] Starting recovery of all containers Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.383378 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.36:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189d299aa1574428 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,LastTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393049 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393149 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393174 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393194 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393211 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393233 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393253 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393274 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393295 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393315 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393334 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393352 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393370 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393423 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393473 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393495 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393515 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393532 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393553 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393575 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393596 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393615 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393633 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393689 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393709 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393726 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393749 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393769 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393787 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393806 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393824 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393854 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393881 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393900 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393917 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393934 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393953 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393970 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.393989 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394015 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394034 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394053 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394071 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394089 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394160 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394181 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394199 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394221 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394241 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394259 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394279 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394296 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394320 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394340 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394363 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394416 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394437 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394457 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394477 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394496 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394515 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394533 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394551 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394575 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394593 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394611 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394630 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394648 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394670 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394688 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394706 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394724 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394746 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394763 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394782 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394801 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394819 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394837 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394855 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394874 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394891 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394909 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394927 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394945 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394964 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.394983 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395063 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395315 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395346 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395365 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395382 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395405 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395424 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395442 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395461 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395480 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395497 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395539 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395557 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395575 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395593 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395611 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395628 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395647 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395673 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395693 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395713 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395731 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395750 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.395769 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398403 4751 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398447 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398470 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398491 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398511 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398531 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398549 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398570 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398589 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398608 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398628 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398646 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398665 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398684 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398701 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398719 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398736 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398756 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398774 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398793 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398813 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398831 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398849 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398870 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398905 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398924 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398942 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398961 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398980 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.398998 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399016 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399034 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399052 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399072 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399091 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399141 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399169 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399187 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399204 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399222 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399241 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399260 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399278 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399295 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399314 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399331 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399351 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399369 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399387 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399405 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399424 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399440 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399461 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399479 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399496 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399550 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399574 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399593 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399609 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399626 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399689 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399709 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399728 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399748 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399765 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399792 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399809 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399827 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399847 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399864 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399882 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399901 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399920 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399937 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399955 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399972 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.399989 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400007 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400022 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400038 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400055 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400789 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400823 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400840 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400862 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400879 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400896 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400916 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400932 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400949 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400965 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.400982 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.401001 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.401018 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.401034 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.401054 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.401074 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.401095 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.401141 4751 reconstruct.go:97] "Volume reconstruction finished" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.401155 4751 reconciler.go:26] "Reconciler: start to sync state" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.413687 4751 manager.go:324] Recovery completed Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.431950 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.434297 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.434364 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.434417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.435492 4751 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.435526 4751 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.435580 4751 state_mem.go:36] "Initialized new in-memory state store" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.449970 4751 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.450289 4751 policy_none.go:49] "None policy: Start" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.451402 4751 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.451450 4751 state_mem.go:35] "Initializing new in-memory state store" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.452391 4751 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.453169 4751 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.453278 4751 kubelet.go:2335] "Starting kubelet main sync loop" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.453395 4751 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.454030 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.454074 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.471797 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.494477 4751 manager.go:334] "Starting Device Plugin manager" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.494917 4751 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.494952 4751 server.go:79] "Starting device plugin registration server" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.495648 4751 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.495681 4751 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.495867 4751 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.495991 4751 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.496004 4751 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.507510 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.554046 4751 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.554233 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.555751 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.555786 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.555797 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.555967 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.556462 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.556513 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.556848 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.556879 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.556919 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557000 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557176 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557231 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557402 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557428 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557437 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557787 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557809 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557818 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.557937 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.558416 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.558458 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.558610 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.558630 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.558637 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.558740 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.558798 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.558817 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.559083 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.559265 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.559332 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.559987 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560011 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560019 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560398 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560429 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560440 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560600 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560665 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560688 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560695 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.560815 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.561596 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.561654 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.561677 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.581991 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="400ms" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.595880 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.596969 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.597128 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.597232 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.597333 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.597882 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.36:6443: connect: connection refused" node="crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605454 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605522 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605565 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605663 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605699 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605718 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605760 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605778 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605795 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605837 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605853 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605870 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605901 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605918 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.605934 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.706692 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.706913 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707006 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707137 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707079 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707227 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707197 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707300 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707329 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707400 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707370 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707498 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707549 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707622 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707647 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707713 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707699 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707784 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707807 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707869 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707894 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707945 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.707978 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.708023 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.708132 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.708043 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.708174 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.708234 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.708242 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.708316 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.798470 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.800845 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.800896 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.800910 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.800946 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.801599 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.36:6443: connect: connection refused" node="crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.881768 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.895799 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.920576 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.929385 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: I0316 00:06:12.934502 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.940016 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-39778b2dcac345027836166a606c2f4b431d0d257edc5582c03922adda90a826 WatchSource:0}: Error finding container 39778b2dcac345027836166a606c2f4b431d0d257edc5582c03922adda90a826: Status 404 returned error can't find the container with id 39778b2dcac345027836166a606c2f4b431d0d257edc5582c03922adda90a826 Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.942916 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b52459bfa55475a570fe5bea0c628a965a900c853961b02c9e32b12de443a3a3 WatchSource:0}: Error finding container b52459bfa55475a570fe5bea0c628a965a900c853961b02c9e32b12de443a3a3: Status 404 returned error can't find the container with id b52459bfa55475a570fe5bea0c628a965a900c853961b02c9e32b12de443a3a3 Mar 16 00:06:12 crc kubenswrapper[4751]: W0316 00:06:12.964448 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-edc17b63d1c14273695b111a6216ea18131577d06fb760c52f1915210f963dc0 WatchSource:0}: Error finding container edc17b63d1c14273695b111a6216ea18131577d06fb760c52f1915210f963dc0: Status 404 returned error can't find the container with id edc17b63d1c14273695b111a6216ea18131577d06fb760c52f1915210f963dc0 Mar 16 00:06:12 crc kubenswrapper[4751]: E0316 00:06:12.983715 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="800ms" Mar 16 00:06:13 crc kubenswrapper[4751]: W0316 00:06:13.198079 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:13 crc kubenswrapper[4751]: E0316 00:06:13.198255 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.202746 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.205087 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.205213 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.205242 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.205290 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:13 crc kubenswrapper[4751]: E0316 00:06:13.205878 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.36:6443: connect: connection refused" node="crc" Mar 16 00:06:13 crc kubenswrapper[4751]: W0316 00:06:13.345180 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:13 crc kubenswrapper[4751]: E0316 00:06:13.345339 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.366034 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:13 crc kubenswrapper[4751]: W0316 00:06:13.372876 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:13 crc kubenswrapper[4751]: E0316 00:06:13.372980 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.459066 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"39778b2dcac345027836166a606c2f4b431d0d257edc5582c03922adda90a826"} Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.460207 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"edc17b63d1c14273695b111a6216ea18131577d06fb760c52f1915210f963dc0"} Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.461568 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ade95c642c00b5abf9c2346d6fb10f83057e6e73b4f04480f1b050a5babe705c"} Mar 16 00:06:13 crc kubenswrapper[4751]: W0316 00:06:13.463000 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:13 crc kubenswrapper[4751]: E0316 00:06:13.463072 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.463124 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3173501f9708740f3701425a6fc27d7f235502fa23366f73666522eb666d7ec9"} Mar 16 00:06:13 crc kubenswrapper[4751]: I0316 00:06:13.464023 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b52459bfa55475a570fe5bea0c628a965a900c853961b02c9e32b12de443a3a3"} Mar 16 00:06:13 crc kubenswrapper[4751]: E0316 00:06:13.785098 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="1.6s" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.006863 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.009312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.009384 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.009406 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.009450 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:14 crc kubenswrapper[4751]: E0316 00:06:14.010200 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.36:6443: connect: connection refused" node="crc" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.365849 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.366875 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 16 00:06:14 crc kubenswrapper[4751]: E0316 00:06:14.368434 4751 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.471793 4751 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="092e62c508251707d69291daf393e5f3c9c4e8b9e63956368b097c04a24eafaa" exitCode=0 Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.471912 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"092e62c508251707d69291daf393e5f3c9c4e8b9e63956368b097c04a24eafaa"} Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.472072 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.473431 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.473501 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.473520 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.482066 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cd295ca6ea9ebd13c67f8732d72d23bd26595233b2b54b939787cac45d122f5a"} Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.482173 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"30486f003ec44598fd2cf7109690d24d0a3d6cedadac353927da4c1d686e82f1"} Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.482184 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.482196 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0f5e4ee6d0e09fec4c1b4fb2c76ed93e5881bd6f2277e4fe3c93fdee6793fd7e"} Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.482409 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eb5cc30849ffd9367bd9dba9e0eccf95ba6e364bee9b1b05ed74e4bb0cdda793"} Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.483371 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.483428 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.483447 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.485602 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057" exitCode=0 Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.485744 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057"} Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.485841 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.487525 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.487576 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.487596 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.488887 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.488919 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ab5dc4bf910d51399e06fc52693ac80f8f08a639bf8d1ec4464a3e18ff31b394"} Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.488728 4751 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ab5dc4bf910d51399e06fc52693ac80f8f08a639bf8d1ec4464a3e18ff31b394" exitCode=0 Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.490383 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.490475 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.490521 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.490545 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.492867 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.492919 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.493024 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.493381 4751 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="c98061b7db4efc2b3bb9f4864fba66269d116a5205b081e3d79b4644526fe382" exitCode=0 Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.493453 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"c98061b7db4efc2b3bb9f4864fba66269d116a5205b081e3d79b4644526fe382"} Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.493634 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.497064 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.497140 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:14 crc kubenswrapper[4751]: I0316 00:06:14.497162 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:14 crc kubenswrapper[4751]: W0316 00:06:14.990283 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:14 crc kubenswrapper[4751]: E0316 00:06:14.990666 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:15 crc kubenswrapper[4751]: W0316 00:06:15.116544 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:15 crc kubenswrapper[4751]: E0316 00:06:15.116652 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.365303 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:15 crc kubenswrapper[4751]: E0316 00:06:15.386315 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="3.2s" Mar 16 00:06:15 crc kubenswrapper[4751]: W0316 00:06:15.457087 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:15 crc kubenswrapper[4751]: E0316 00:06:15.457210 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.498860 4751 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9eb24956c13ecb8e064d566f98720cde2cfe8dfc65bf0e2c940653b4c07ae281" exitCode=0 Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.498935 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9eb24956c13ecb8e064d566f98720cde2cfe8dfc65bf0e2c940653b4c07ae281"} Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.498983 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.499930 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.499963 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.499975 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.505072 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.505056 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2ad59fd37cb81fd27b63b4daa4fd13a86df5fec07bdd43b90d73083a4628fc73"} Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.506563 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.506617 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.506636 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.509070 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5a576b19216f2fdb367318a6cf4ca15cda914454c90107e6dbb9b71d6d002fe0"} Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.509129 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"604d9f1b3017201ae2101b09f3d55adafc8eb5c4009c194058185e936b752878"} Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.509141 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b619573bc26ac8e3ce86ce45a9e083027dedfdbc1e653674184f62ac2a180963"} Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.509269 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.510297 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.510320 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.510329 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.513856 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.514160 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a"} Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.514193 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8"} Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.514205 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d"} Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.514213 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a"} Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.514772 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.514802 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.514811 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.611268 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.613539 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.613587 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.613600 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:15 crc kubenswrapper[4751]: I0316 00:06:15.613631 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:15 crc kubenswrapper[4751]: E0316 00:06:15.614233 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.36:6443: connect: connection refused" node="crc" Mar 16 00:06:15 crc kubenswrapper[4751]: W0316 00:06:15.740125 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.36:6443: connect: connection refused Mar 16 00:06:15 crc kubenswrapper[4751]: E0316 00:06:15.740232 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.36:6443: connect: connection refused" logger="UnhandledError" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.521505 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6629332d01ca0fc8b0fa4fa25bc9b3f001a4e085e4f7865fd8d0da4fd4bf8523"} Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.521665 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.524147 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.524224 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.524247 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.524948 4751 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5eaa0195915ad30a72a368bd3833d5f43dff01d54df34ecab0438392f49e91e1" exitCode=0 Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.525092 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5eaa0195915ad30a72a368bd3833d5f43dff01d54df34ecab0438392f49e91e1"} Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.525212 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.525218 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.526293 4751 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.526414 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.526803 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.526853 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.526871 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.526887 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.526930 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.526949 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.527885 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.527917 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.527936 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:16 crc kubenswrapper[4751]: I0316 00:06:16.657967 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:17 crc kubenswrapper[4751]: I0316 00:06:17.537604 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5609b66b8751522544f0c1dd6edb553eb59f88af40f01cb1c0caab5d242d6666"} Mar 16 00:06:17 crc kubenswrapper[4751]: I0316 00:06:17.538542 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:17 crc kubenswrapper[4751]: I0316 00:06:17.538702 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"873102e15f1366da0187ed08a6751564f10cace5a8c1f2edd0de7501e16ccbcc"} Mar 16 00:06:17 crc kubenswrapper[4751]: I0316 00:06:17.538855 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"843e9d1a62d2486dc79bae6712c8c409fae210bbfb33a656d899fcbd9c5bb0c2"} Mar 16 00:06:17 crc kubenswrapper[4751]: I0316 00:06:17.537768 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:17 crc kubenswrapper[4751]: I0316 00:06:17.541071 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:17 crc kubenswrapper[4751]: I0316 00:06:17.541155 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:17 crc kubenswrapper[4751]: I0316 00:06:17.541174 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.548606 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4634b7be5c7bfe0eaea4d2ba85e93f6b6331a0408874555c4b91fb412d53f525"} Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.548673 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.548692 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f9c34d073cc679b577714c5a7ab3a1b3008a9b8e682fc77435188b80749727a6"} Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.548673 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.550228 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.550261 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.550303 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.550324 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.550276 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.550386 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.753171 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.815050 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.816906 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.816961 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.816978 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.817013 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.830917 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.831211 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.832544 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.832588 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:18 crc kubenswrapper[4751]: I0316 00:06:18.832606 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.408000 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.552326 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.552326 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.553955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.554002 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.554031 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.554067 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.554037 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.554390 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.801288 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.801566 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.803229 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.803298 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:19 crc kubenswrapper[4751]: I0316 00:06:19.803322 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:21 crc kubenswrapper[4751]: I0316 00:06:21.848590 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 16 00:06:21 crc kubenswrapper[4751]: I0316 00:06:21.848853 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:21 crc kubenswrapper[4751]: I0316 00:06:21.850628 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:21 crc kubenswrapper[4751]: I0316 00:06:21.850684 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:21 crc kubenswrapper[4751]: I0316 00:06:21.850701 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:22 crc kubenswrapper[4751]: I0316 00:06:22.130085 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 16 00:06:22 crc kubenswrapper[4751]: E0316 00:06:22.507735 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:06:22 crc kubenswrapper[4751]: I0316 00:06:22.561884 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:22 crc kubenswrapper[4751]: I0316 00:06:22.563151 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:22 crc kubenswrapper[4751]: I0316 00:06:22.563241 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:22 crc kubenswrapper[4751]: I0316 00:06:22.563258 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:22 crc kubenswrapper[4751]: I0316 00:06:22.802266 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 16 00:06:22 crc kubenswrapper[4751]: I0316 00:06:22.802423 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.070500 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.070816 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.072588 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.072656 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.072684 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.124861 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.135180 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.469417 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.564586 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.565887 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.565944 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.565957 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:23 crc kubenswrapper[4751]: I0316 00:06:23.571219 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:24 crc kubenswrapper[4751]: I0316 00:06:24.567647 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:24 crc kubenswrapper[4751]: I0316 00:06:24.569039 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:24 crc kubenswrapper[4751]: I0316 00:06:24.569093 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:24 crc kubenswrapper[4751]: I0316 00:06:24.569131 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:25 crc kubenswrapper[4751]: I0316 00:06:25.570296 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:25 crc kubenswrapper[4751]: I0316 00:06:25.571555 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:25 crc kubenswrapper[4751]: I0316 00:06:25.571601 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:25 crc kubenswrapper[4751]: I0316 00:06:25.571612 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:26 crc kubenswrapper[4751]: I0316 00:06:26.367235 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 16 00:06:26 crc kubenswrapper[4751]: E0316 00:06:26.528251 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 16 00:06:26 crc kubenswrapper[4751]: W0316 00:06:26.529516 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z Mar 16 00:06:26 crc kubenswrapper[4751]: E0316 00:06:26.529637 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:26 crc kubenswrapper[4751]: W0316 00:06:26.531301 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z Mar 16 00:06:26 crc kubenswrapper[4751]: E0316 00:06:26.531396 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:26 crc kubenswrapper[4751]: E0316 00:06:26.534690 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189d299aa1574428 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,LastTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:06:26 crc kubenswrapper[4751]: W0316 00:06:26.537330 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z Mar 16 00:06:26 crc kubenswrapper[4751]: E0316 00:06:26.537463 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:26 crc kubenswrapper[4751]: I0316 00:06:26.537812 4751 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 16 00:06:26 crc kubenswrapper[4751]: I0316 00:06:26.537901 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 16 00:06:26 crc kubenswrapper[4751]: E0316 00:06:26.538772 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z" node="crc" Mar 16 00:06:26 crc kubenswrapper[4751]: W0316 00:06:26.542098 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z Mar 16 00:06:26 crc kubenswrapper[4751]: E0316 00:06:26.542249 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:26 crc kubenswrapper[4751]: I0316 00:06:26.543038 4751 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 16 00:06:26 crc kubenswrapper[4751]: I0316 00:06:26.543166 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 16 00:06:26 crc kubenswrapper[4751]: E0316 00:06:26.544713 4751 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:27 crc kubenswrapper[4751]: I0316 00:06:27.369620 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:27Z is after 2026-02-23T05:33:13Z Mar 16 00:06:27 crc kubenswrapper[4751]: I0316 00:06:27.578316 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 16 00:06:27 crc kubenswrapper[4751]: I0316 00:06:27.580657 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6629332d01ca0fc8b0fa4fa25bc9b3f001a4e085e4f7865fd8d0da4fd4bf8523" exitCode=255 Mar 16 00:06:27 crc kubenswrapper[4751]: I0316 00:06:27.580718 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6629332d01ca0fc8b0fa4fa25bc9b3f001a4e085e4f7865fd8d0da4fd4bf8523"} Mar 16 00:06:27 crc kubenswrapper[4751]: I0316 00:06:27.581009 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:27 crc kubenswrapper[4751]: I0316 00:06:27.582280 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:27 crc kubenswrapper[4751]: I0316 00:06:27.582354 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:27 crc kubenswrapper[4751]: I0316 00:06:27.582368 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:27 crc kubenswrapper[4751]: I0316 00:06:27.583456 4751 scope.go:117] "RemoveContainer" containerID="6629332d01ca0fc8b0fa4fa25bc9b3f001a4e085e4f7865fd8d0da4fd4bf8523" Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.369851 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:28Z is after 2026-02-23T05:33:13Z Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.590842 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.591851 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.594879 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4385f9ba571fd0de40eede9229fc37366c7cd5de16b514e047dbdbb00603d7ac" exitCode=255 Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.594952 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4385f9ba571fd0de40eede9229fc37366c7cd5de16b514e047dbdbb00603d7ac"} Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.595019 4751 scope.go:117] "RemoveContainer" containerID="6629332d01ca0fc8b0fa4fa25bc9b3f001a4e085e4f7865fd8d0da4fd4bf8523" Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.595233 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.596680 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.596733 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.596755 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:28 crc kubenswrapper[4751]: I0316 00:06:28.597522 4751 scope.go:117] "RemoveContainer" containerID="4385f9ba571fd0de40eede9229fc37366c7cd5de16b514e047dbdbb00603d7ac" Mar 16 00:06:28 crc kubenswrapper[4751]: E0316 00:06:28.597882 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:06:29 crc kubenswrapper[4751]: I0316 00:06:29.370684 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:29Z is after 2026-02-23T05:33:13Z Mar 16 00:06:29 crc kubenswrapper[4751]: I0316 00:06:29.417335 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:29 crc kubenswrapper[4751]: I0316 00:06:29.600519 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 16 00:06:29 crc kubenswrapper[4751]: I0316 00:06:29.603712 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:29 crc kubenswrapper[4751]: I0316 00:06:29.605606 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:29 crc kubenswrapper[4751]: I0316 00:06:29.605714 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:29 crc kubenswrapper[4751]: I0316 00:06:29.605736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:29 crc kubenswrapper[4751]: I0316 00:06:29.607331 4751 scope.go:117] "RemoveContainer" containerID="4385f9ba571fd0de40eede9229fc37366c7cd5de16b514e047dbdbb00603d7ac" Mar 16 00:06:29 crc kubenswrapper[4751]: E0316 00:06:29.607890 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:06:29 crc kubenswrapper[4751]: I0316 00:06:29.611087 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:30 crc kubenswrapper[4751]: I0316 00:06:30.369811 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:30Z is after 2026-02-23T05:33:13Z Mar 16 00:06:30 crc kubenswrapper[4751]: I0316 00:06:30.607189 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:30 crc kubenswrapper[4751]: I0316 00:06:30.608571 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:30 crc kubenswrapper[4751]: I0316 00:06:30.608611 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:30 crc kubenswrapper[4751]: I0316 00:06:30.608623 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:30 crc kubenswrapper[4751]: I0316 00:06:30.609227 4751 scope.go:117] "RemoveContainer" containerID="4385f9ba571fd0de40eede9229fc37366c7cd5de16b514e047dbdbb00603d7ac" Mar 16 00:06:30 crc kubenswrapper[4751]: E0316 00:06:30.609443 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:06:31 crc kubenswrapper[4751]: I0316 00:06:31.369782 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:31Z is after 2026-02-23T05:33:13Z Mar 16 00:06:31 crc kubenswrapper[4751]: I0316 00:06:31.885046 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 16 00:06:31 crc kubenswrapper[4751]: I0316 00:06:31.885296 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:31 crc kubenswrapper[4751]: I0316 00:06:31.886923 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:31 crc kubenswrapper[4751]: I0316 00:06:31.886959 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:31 crc kubenswrapper[4751]: I0316 00:06:31.886970 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:31 crc kubenswrapper[4751]: I0316 00:06:31.906047 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.370830 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:32Z is after 2026-02-23T05:33:13Z Mar 16 00:06:32 crc kubenswrapper[4751]: E0316 00:06:32.507899 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.613413 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.615139 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.615191 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.615210 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.802457 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.802595 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 16 00:06:32 crc kubenswrapper[4751]: E0316 00:06:32.934411 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:32Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.939498 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.940895 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.940959 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.940978 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:32 crc kubenswrapper[4751]: I0316 00:06:32.941014 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:32 crc kubenswrapper[4751]: E0316 00:06:32.945966 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:32Z is after 2026-02-23T05:33:13Z" node="crc" Mar 16 00:06:33 crc kubenswrapper[4751]: I0316 00:06:33.370397 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:33Z is after 2026-02-23T05:33:13Z Mar 16 00:06:33 crc kubenswrapper[4751]: W0316 00:06:33.859147 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:33Z is after 2026-02-23T05:33:13Z Mar 16 00:06:33 crc kubenswrapper[4751]: E0316 00:06:33.859289 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:34 crc kubenswrapper[4751]: W0316 00:06:34.137543 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:34Z is after 2026-02-23T05:33:13Z Mar 16 00:06:34 crc kubenswrapper[4751]: E0316 00:06:34.137662 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:34Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.369745 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:34Z is after 2026-02-23T05:33:13Z Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.501993 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.502248 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.503551 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.503615 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.503636 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.504521 4751 scope.go:117] "RemoveContainer" containerID="4385f9ba571fd0de40eede9229fc37366c7cd5de16b514e047dbdbb00603d7ac" Mar 16 00:06:34 crc kubenswrapper[4751]: E0316 00:06:34.504863 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.826846 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.827158 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.828812 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.828915 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.828941 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.829952 4751 scope.go:117] "RemoveContainer" containerID="4385f9ba571fd0de40eede9229fc37366c7cd5de16b514e047dbdbb00603d7ac" Mar 16 00:06:34 crc kubenswrapper[4751]: E0316 00:06:34.830290 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:06:34 crc kubenswrapper[4751]: I0316 00:06:34.978134 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 16 00:06:34 crc kubenswrapper[4751]: E0316 00:06:34.984157 4751 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:34Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:35 crc kubenswrapper[4751]: W0316 00:06:35.342389 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:35Z is after 2026-02-23T05:33:13Z Mar 16 00:06:35 crc kubenswrapper[4751]: E0316 00:06:35.342534 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:35Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:35 crc kubenswrapper[4751]: I0316 00:06:35.369639 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:35Z is after 2026-02-23T05:33:13Z Mar 16 00:06:36 crc kubenswrapper[4751]: I0316 00:06:36.369830 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:36Z is after 2026-02-23T05:33:13Z Mar 16 00:06:36 crc kubenswrapper[4751]: E0316 00:06:36.539955 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:36Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189d299aa1574428 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,LastTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:06:37 crc kubenswrapper[4751]: I0316 00:06:37.370166 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:37Z is after 2026-02-23T05:33:13Z Mar 16 00:06:38 crc kubenswrapper[4751]: I0316 00:06:38.368475 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:38Z is after 2026-02-23T05:33:13Z Mar 16 00:06:38 crc kubenswrapper[4751]: W0316 00:06:38.797703 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:38Z is after 2026-02-23T05:33:13Z Mar 16 00:06:38 crc kubenswrapper[4751]: E0316 00:06:38.797859 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:38Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:39 crc kubenswrapper[4751]: I0316 00:06:39.370561 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:39Z is after 2026-02-23T05:33:13Z Mar 16 00:06:39 crc kubenswrapper[4751]: E0316 00:06:39.940747 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:39Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 16 00:06:39 crc kubenswrapper[4751]: I0316 00:06:39.946996 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:39 crc kubenswrapper[4751]: I0316 00:06:39.948850 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:39 crc kubenswrapper[4751]: I0316 00:06:39.949054 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:39 crc kubenswrapper[4751]: I0316 00:06:39.949258 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:39 crc kubenswrapper[4751]: I0316 00:06:39.949433 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:39 crc kubenswrapper[4751]: E0316 00:06:39.954646 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:39Z is after 2026-02-23T05:33:13Z" node="crc" Mar 16 00:06:40 crc kubenswrapper[4751]: I0316 00:06:40.371465 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:40Z is after 2026-02-23T05:33:13Z Mar 16 00:06:41 crc kubenswrapper[4751]: I0316 00:06:41.370781 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:41Z is after 2026-02-23T05:33:13Z Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.374737 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:42Z is after 2026-02-23T05:33:13Z Mar 16 00:06:42 crc kubenswrapper[4751]: E0316 00:06:42.508088 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.802395 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.802528 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.802623 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.802833 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.804491 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.804559 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.804577 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.805267 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"0f5e4ee6d0e09fec4c1b4fb2c76ed93e5881bd6f2277e4fe3c93fdee6793fd7e"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 16 00:06:42 crc kubenswrapper[4751]: I0316 00:06:42.805542 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://0f5e4ee6d0e09fec4c1b4fb2c76ed93e5881bd6f2277e4fe3c93fdee6793fd7e" gracePeriod=30 Mar 16 00:06:43 crc kubenswrapper[4751]: I0316 00:06:43.370357 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:43Z is after 2026-02-23T05:33:13Z Mar 16 00:06:43 crc kubenswrapper[4751]: I0316 00:06:43.650286 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 16 00:06:43 crc kubenswrapper[4751]: I0316 00:06:43.651254 4751 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0f5e4ee6d0e09fec4c1b4fb2c76ed93e5881bd6f2277e4fe3c93fdee6793fd7e" exitCode=255 Mar 16 00:06:43 crc kubenswrapper[4751]: I0316 00:06:43.651301 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0f5e4ee6d0e09fec4c1b4fb2c76ed93e5881bd6f2277e4fe3c93fdee6793fd7e"} Mar 16 00:06:43 crc kubenswrapper[4751]: I0316 00:06:43.651350 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"36c6b8a3de091f965af748d7e72d167e3be7db4afa2ca3a554b0ec2e402fb4cf"} Mar 16 00:06:43 crc kubenswrapper[4751]: I0316 00:06:43.651465 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:43 crc kubenswrapper[4751]: I0316 00:06:43.652664 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:43 crc kubenswrapper[4751]: I0316 00:06:43.652689 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:43 crc kubenswrapper[4751]: I0316 00:06:43.652699 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:44 crc kubenswrapper[4751]: I0316 00:06:44.370558 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:44Z is after 2026-02-23T05:33:13Z Mar 16 00:06:45 crc kubenswrapper[4751]: I0316 00:06:45.371156 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:45Z is after 2026-02-23T05:33:13Z Mar 16 00:06:46 crc kubenswrapper[4751]: I0316 00:06:46.370201 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:46Z is after 2026-02-23T05:33:13Z Mar 16 00:06:46 crc kubenswrapper[4751]: E0316 00:06:46.545408 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:46Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189d299aa1574428 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,LastTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:06:46 crc kubenswrapper[4751]: E0316 00:06:46.948302 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:46Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 16 00:06:46 crc kubenswrapper[4751]: I0316 00:06:46.955127 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:46 crc kubenswrapper[4751]: I0316 00:06:46.957033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:46 crc kubenswrapper[4751]: I0316 00:06:46.957164 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:46 crc kubenswrapper[4751]: I0316 00:06:46.957188 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:46 crc kubenswrapper[4751]: I0316 00:06:46.957234 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:46 crc kubenswrapper[4751]: E0316 00:06:46.962723 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:46Z is after 2026-02-23T05:33:13Z" node="crc" Mar 16 00:06:47 crc kubenswrapper[4751]: I0316 00:06:47.369974 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:47Z is after 2026-02-23T05:33:13Z Mar 16 00:06:47 crc kubenswrapper[4751]: I0316 00:06:47.454704 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:47 crc kubenswrapper[4751]: I0316 00:06:47.456945 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:47 crc kubenswrapper[4751]: I0316 00:06:47.457012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:47 crc kubenswrapper[4751]: I0316 00:06:47.457036 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:47 crc kubenswrapper[4751]: I0316 00:06:47.458013 4751 scope.go:117] "RemoveContainer" containerID="4385f9ba571fd0de40eede9229fc37366c7cd5de16b514e047dbdbb00603d7ac" Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.369792 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:48Z is after 2026-02-23T05:33:13Z Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.668550 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.669640 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.672247 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c7331ac777f3e8c653bab9d0bae9dc6e67a6c2d68ba440c06eb083a82867ca3a" exitCode=255 Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.672308 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c7331ac777f3e8c653bab9d0bae9dc6e67a6c2d68ba440c06eb083a82867ca3a"} Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.672471 4751 scope.go:117] "RemoveContainer" containerID="4385f9ba571fd0de40eede9229fc37366c7cd5de16b514e047dbdbb00603d7ac" Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.672650 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.674039 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.674095 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.674184 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:48 crc kubenswrapper[4751]: I0316 00:06:48.680316 4751 scope.go:117] "RemoveContainer" containerID="c7331ac777f3e8c653bab9d0bae9dc6e67a6c2d68ba440c06eb083a82867ca3a" Mar 16 00:06:48 crc kubenswrapper[4751]: E0316 00:06:48.680615 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:06:49 crc kubenswrapper[4751]: I0316 00:06:49.370386 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:49Z is after 2026-02-23T05:33:13Z Mar 16 00:06:49 crc kubenswrapper[4751]: I0316 00:06:49.678746 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 16 00:06:49 crc kubenswrapper[4751]: I0316 00:06:49.801584 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:49 crc kubenswrapper[4751]: I0316 00:06:49.801779 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:49 crc kubenswrapper[4751]: I0316 00:06:49.803341 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:49 crc kubenswrapper[4751]: I0316 00:06:49.803418 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:49 crc kubenswrapper[4751]: I0316 00:06:49.803444 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:50 crc kubenswrapper[4751]: I0316 00:06:50.370574 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:50Z is after 2026-02-23T05:33:13Z Mar 16 00:06:51 crc kubenswrapper[4751]: I0316 00:06:51.370512 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:51Z is after 2026-02-23T05:33:13Z Mar 16 00:06:52 crc kubenswrapper[4751]: I0316 00:06:52.370569 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:52Z is after 2026-02-23T05:33:13Z Mar 16 00:06:52 crc kubenswrapper[4751]: E0316 00:06:52.508800 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:06:52 crc kubenswrapper[4751]: I0316 00:06:52.572982 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 16 00:06:52 crc kubenswrapper[4751]: E0316 00:06:52.579267 4751 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:52Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:52 crc kubenswrapper[4751]: E0316 00:06:52.580390 4751 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 16 00:06:52 crc kubenswrapper[4751]: I0316 00:06:52.802092 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 16 00:06:52 crc kubenswrapper[4751]: I0316 00:06:52.802217 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 16 00:06:53 crc kubenswrapper[4751]: W0316 00:06:53.291060 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:53Z is after 2026-02-23T05:33:13Z Mar 16 00:06:53 crc kubenswrapper[4751]: E0316 00:06:53.291237 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:53Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.370616 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:53Z is after 2026-02-23T05:33:13Z Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.468763 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.469055 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.471251 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.471335 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.471358 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:53 crc kubenswrapper[4751]: E0316 00:06:53.954475 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:53Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.963883 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.965825 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.965901 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.965927 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:53 crc kubenswrapper[4751]: I0316 00:06:53.965985 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:06:53 crc kubenswrapper[4751]: E0316 00:06:53.970865 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:53Z is after 2026-02-23T05:33:13Z" node="crc" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.371528 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:54Z is after 2026-02-23T05:33:13Z Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.502164 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.502469 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.504252 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.504298 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.504319 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.505209 4751 scope.go:117] "RemoveContainer" containerID="c7331ac777f3e8c653bab9d0bae9dc6e67a6c2d68ba440c06eb083a82867ca3a" Mar 16 00:06:54 crc kubenswrapper[4751]: E0316 00:06:54.505536 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.827062 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.827386 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.829374 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.829473 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.829495 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:06:54 crc kubenswrapper[4751]: I0316 00:06:54.833520 4751 scope.go:117] "RemoveContainer" containerID="c7331ac777f3e8c653bab9d0bae9dc6e67a6c2d68ba440c06eb083a82867ca3a" Mar 16 00:06:54 crc kubenswrapper[4751]: E0316 00:06:54.834000 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:06:55 crc kubenswrapper[4751]: I0316 00:06:55.370596 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:55Z is after 2026-02-23T05:33:13Z Mar 16 00:06:55 crc kubenswrapper[4751]: W0316 00:06:55.893638 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:55Z is after 2026-02-23T05:33:13Z Mar 16 00:06:55 crc kubenswrapper[4751]: E0316 00:06:55.893760 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:56 crc kubenswrapper[4751]: I0316 00:06:56.370643 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:56Z is after 2026-02-23T05:33:13Z Mar 16 00:06:56 crc kubenswrapper[4751]: E0316 00:06:56.550397 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:56Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189d299aa1574428 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,LastTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:06:57 crc kubenswrapper[4751]: I0316 00:06:57.371099 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:57Z is after 2026-02-23T05:33:13Z Mar 16 00:06:58 crc kubenswrapper[4751]: I0316 00:06:58.369940 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:58Z is after 2026-02-23T05:33:13Z Mar 16 00:06:58 crc kubenswrapper[4751]: W0316 00:06:58.416402 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:58Z is after 2026-02-23T05:33:13Z Mar 16 00:06:58 crc kubenswrapper[4751]: E0316 00:06:58.416535 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:58Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:06:59 crc kubenswrapper[4751]: I0316 00:06:59.370562 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:06:59Z is after 2026-02-23T05:33:13Z Mar 16 00:07:00 crc kubenswrapper[4751]: I0316 00:07:00.371289 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:00Z is after 2026-02-23T05:33:13Z Mar 16 00:07:00 crc kubenswrapper[4751]: W0316 00:07:00.375968 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:00Z is after 2026-02-23T05:33:13Z Mar 16 00:07:00 crc kubenswrapper[4751]: E0316 00:07:00.376132 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 16 00:07:00 crc kubenswrapper[4751]: E0316 00:07:00.959940 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:00Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 16 00:07:00 crc kubenswrapper[4751]: I0316 00:07:00.972046 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:00 crc kubenswrapper[4751]: I0316 00:07:00.974015 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:00 crc kubenswrapper[4751]: I0316 00:07:00.974277 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:00 crc kubenswrapper[4751]: I0316 00:07:00.974459 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:00 crc kubenswrapper[4751]: I0316 00:07:00.974652 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:07:00 crc kubenswrapper[4751]: E0316 00:07:00.979887 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:00Z is after 2026-02-23T05:33:13Z" node="crc" Mar 16 00:07:01 crc kubenswrapper[4751]: I0316 00:07:01.370365 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:01Z is after 2026-02-23T05:33:13Z Mar 16 00:07:02 crc kubenswrapper[4751]: I0316 00:07:02.370542 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:02Z is after 2026-02-23T05:33:13Z Mar 16 00:07:02 crc kubenswrapper[4751]: E0316 00:07:02.509404 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:07:02 crc kubenswrapper[4751]: I0316 00:07:02.802586 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 16 00:07:02 crc kubenswrapper[4751]: I0316 00:07:02.802696 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 16 00:07:03 crc kubenswrapper[4751]: I0316 00:07:03.371596 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:03Z is after 2026-02-23T05:33:13Z Mar 16 00:07:04 crc kubenswrapper[4751]: I0316 00:07:04.370337 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:04Z is after 2026-02-23T05:33:13Z Mar 16 00:07:05 crc kubenswrapper[4751]: I0316 00:07:05.370168 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:05Z is after 2026-02-23T05:33:13Z Mar 16 00:07:06 crc kubenswrapper[4751]: I0316 00:07:06.371359 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:06Z is after 2026-02-23T05:33:13Z Mar 16 00:07:06 crc kubenswrapper[4751]: E0316 00:07:06.555031 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:06Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189d299aa1574428 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,LastTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.368823 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:07Z is after 2026-02-23T05:33:13Z Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.454400 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.456296 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.456340 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.456354 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.457213 4751 scope.go:117] "RemoveContainer" containerID="c7331ac777f3e8c653bab9d0bae9dc6e67a6c2d68ba440c06eb083a82867ca3a" Mar 16 00:07:07 crc kubenswrapper[4751]: E0316 00:07:07.457496 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:07:07 crc kubenswrapper[4751]: E0316 00:07:07.966753 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:07Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.980873 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.982669 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.982731 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.982751 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:07 crc kubenswrapper[4751]: I0316 00:07:07.982791 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:07:07 crc kubenswrapper[4751]: E0316 00:07:07.987324 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:07Z is after 2026-02-23T05:33:13Z" node="crc" Mar 16 00:07:08 crc kubenswrapper[4751]: I0316 00:07:08.370156 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:08Z is after 2026-02-23T05:33:13Z Mar 16 00:07:08 crc kubenswrapper[4751]: I0316 00:07:08.836772 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 16 00:07:08 crc kubenswrapper[4751]: I0316 00:07:08.837647 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:08 crc kubenswrapper[4751]: I0316 00:07:08.839022 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:08 crc kubenswrapper[4751]: I0316 00:07:08.839094 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:08 crc kubenswrapper[4751]: I0316 00:07:08.839157 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:09 crc kubenswrapper[4751]: I0316 00:07:09.370903 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:09Z is after 2026-02-23T05:33:13Z Mar 16 00:07:10 crc kubenswrapper[4751]: I0316 00:07:10.371029 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:10Z is after 2026-02-23T05:33:13Z Mar 16 00:07:11 crc kubenswrapper[4751]: I0316 00:07:11.370696 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:11Z is after 2026-02-23T05:33:13Z Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.367994 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:12Z is after 2026-02-23T05:33:13Z Mar 16 00:07:12 crc kubenswrapper[4751]: E0316 00:07:12.510006 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.802771 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.802894 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.802999 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.803300 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.806822 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.806880 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.806906 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.809027 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"36c6b8a3de091f965af748d7e72d167e3be7db4afa2ca3a554b0ec2e402fb4cf"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 16 00:07:12 crc kubenswrapper[4751]: I0316 00:07:12.809351 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://36c6b8a3de091f965af748d7e72d167e3be7db4afa2ca3a554b0ec2e402fb4cf" gracePeriod=30 Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.371920 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:13Z is after 2026-02-23T05:33:13Z Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.764997 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.767054 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.767864 4751 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="36c6b8a3de091f965af748d7e72d167e3be7db4afa2ca3a554b0ec2e402fb4cf" exitCode=255 Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.767957 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"36c6b8a3de091f965af748d7e72d167e3be7db4afa2ca3a554b0ec2e402fb4cf"} Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.768051 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"24e25b434223e4354a4d3853c6c3545a545b93ee1e6b63ea06b828c2591bc080"} Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.768131 4751 scope.go:117] "RemoveContainer" containerID="0f5e4ee6d0e09fec4c1b4fb2c76ed93e5881bd6f2277e4fe3c93fdee6793fd7e" Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.768263 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.769639 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.769691 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:13 crc kubenswrapper[4751]: I0316 00:07:13.769709 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.371040 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:14Z is after 2026-02-23T05:33:13Z Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.773502 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.775208 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.776428 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.776476 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.776498 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:14 crc kubenswrapper[4751]: E0316 00:07:14.974956 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.988064 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.989589 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.989640 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.989660 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:14 crc kubenswrapper[4751]: I0316 00:07:14.989698 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:07:14 crc kubenswrapper[4751]: E0316 00:07:14.996235 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 16 00:07:15 crc kubenswrapper[4751]: I0316 00:07:15.371926 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:16 crc kubenswrapper[4751]: I0316 00:07:16.373478 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.563757 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa1574428 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,LastTimestamp:2026-03-16 00:06:12.362396712 +0000 UTC m=+0.631729032,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.574420 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a12576 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434347382 +0000 UTC m=+0.703679712,LastTimestamp:2026-03-16 00:06:12.434347382 +0000 UTC m=+0.703679712,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.583056 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a20374 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434404212 +0000 UTC m=+0.703736542,LastTimestamp:2026-03-16 00:06:12.434404212 +0000 UTC m=+0.703736542,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.589512 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a2674b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434429771 +0000 UTC m=+0.703762101,LastTimestamp:2026-03-16 00:06:12.434429771 +0000 UTC m=+0.703762101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.597735 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa99615fb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.500731387 +0000 UTC m=+0.770063697,LastTimestamp:2026-03-16 00:06:12.500731387 +0000 UTC m=+0.770063697,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.604799 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a12576\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a12576 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434347382 +0000 UTC m=+0.703679712,LastTimestamp:2026-03-16 00:06:12.555777353 +0000 UTC m=+0.825109653,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.609713 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a20374\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a20374 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434404212 +0000 UTC m=+0.703736542,LastTimestamp:2026-03-16 00:06:12.555792573 +0000 UTC m=+0.825124873,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.614185 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a2674b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a2674b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434429771 +0000 UTC m=+0.703762101,LastTimestamp:2026-03-16 00:06:12.555803633 +0000 UTC m=+0.825135933,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.618535 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a12576\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a12576 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434347382 +0000 UTC m=+0.703679712,LastTimestamp:2026-03-16 00:06:12.556867734 +0000 UTC m=+0.826200034,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.622796 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a20374\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a20374 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434404212 +0000 UTC m=+0.703736542,LastTimestamp:2026-03-16 00:06:12.556885854 +0000 UTC m=+0.826218154,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.626098 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a2674b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a2674b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434429771 +0000 UTC m=+0.703762101,LastTimestamp:2026-03-16 00:06:12.556926774 +0000 UTC m=+0.826259074,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.630236 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a12576\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a12576 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434347382 +0000 UTC m=+0.703679712,LastTimestamp:2026-03-16 00:06:12.55742146 +0000 UTC m=+0.826753750,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.634217 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a20374\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a20374 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434404212 +0000 UTC m=+0.703736542,LastTimestamp:2026-03-16 00:06:12.55743379 +0000 UTC m=+0.826766070,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.638973 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a2674b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a2674b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434429771 +0000 UTC m=+0.703762101,LastTimestamp:2026-03-16 00:06:12.55744164 +0000 UTC m=+0.826773930,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.643177 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a12576\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a12576 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434347382 +0000 UTC m=+0.703679712,LastTimestamp:2026-03-16 00:06:12.557799747 +0000 UTC m=+0.827132037,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.647433 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a20374\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a20374 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434404212 +0000 UTC m=+0.703736542,LastTimestamp:2026-03-16 00:06:12.557814577 +0000 UTC m=+0.827146857,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.653793 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a2674b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a2674b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434429771 +0000 UTC m=+0.703762101,LastTimestamp:2026-03-16 00:06:12.557822797 +0000 UTC m=+0.827155087,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.659087 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a12576\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a12576 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434347382 +0000 UTC m=+0.703679712,LastTimestamp:2026-03-16 00:06:12.55861762 +0000 UTC m=+0.827949910,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.663170 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a20374\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a20374 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434404212 +0000 UTC m=+0.703736542,LastTimestamp:2026-03-16 00:06:12.55863479 +0000 UTC m=+0.827967080,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.667541 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a2674b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a2674b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434429771 +0000 UTC m=+0.703762101,LastTimestamp:2026-03-16 00:06:12.55864151 +0000 UTC m=+0.827973800,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.671925 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a12576\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a12576 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434347382 +0000 UTC m=+0.703679712,LastTimestamp:2026-03-16 00:06:12.558773609 +0000 UTC m=+0.828105939,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.675933 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a20374\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a20374 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434404212 +0000 UTC m=+0.703736542,LastTimestamp:2026-03-16 00:06:12.558809369 +0000 UTC m=+0.828141699,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.679294 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a2674b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a2674b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434429771 +0000 UTC m=+0.703762101,LastTimestamp:2026-03-16 00:06:12.558827539 +0000 UTC m=+0.828159869,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.685859 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a12576\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a12576 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434347382 +0000 UTC m=+0.703679712,LastTimestamp:2026-03-16 00:06:12.560000349 +0000 UTC m=+0.829332639,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.692600 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d299aa5a20374\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d299aa5a20374 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.434404212 +0000 UTC m=+0.703736542,LastTimestamp:2026-03-16 00:06:12.560016649 +0000 UTC m=+0.829348939,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.699071 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299ac48e4fea openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.953206762 +0000 UTC m=+1.222539062,LastTimestamp:2026-03-16 00:06:12.953206762 +0000 UTC m=+1.222539062,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.703811 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299ac49096db openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.953355995 +0000 UTC m=+1.222688285,LastTimestamp:2026-03-16 00:06:12.953355995 +0000 UTC m=+1.222688285,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.710256 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d299ac4bffb56 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.95646191 +0000 UTC m=+1.225794220,LastTimestamp:2026-03-16 00:06:12.95646191 +0000 UTC m=+1.225794220,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.716306 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299ac4e23437 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.958704695 +0000 UTC m=+1.228037005,LastTimestamp:2026-03-16 00:06:12.958704695 +0000 UTC m=+1.228037005,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.719783 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299ac5661dd0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:12.967349712 +0000 UTC m=+1.236682012,LastTimestamp:2026-03-16 00:06:12.967349712 +0000 UTC m=+1.236682012,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.723698 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299ae79647b5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.540931509 +0000 UTC m=+1.810263839,LastTimestamp:2026-03-16 00:06:13.540931509 +0000 UTC m=+1.810263839,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.727037 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299ae833536a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.551223658 +0000 UTC m=+1.820555998,LastTimestamp:2026-03-16 00:06:13.551223658 +0000 UTC m=+1.820555998,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.733791 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299ae8465110 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.55246824 +0000 UTC m=+1.821800520,LastTimestamp:2026-03-16 00:06:13.55246824 +0000 UTC m=+1.821800520,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.737801 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299ae88bc9e5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.557021157 +0000 UTC m=+1.826353497,LastTimestamp:2026-03-16 00:06:13.557021157 +0000 UTC m=+1.826353497,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.743396 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299ae952c7d0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.570062288 +0000 UTC m=+1.839394578,LastTimestamp:2026-03-16 00:06:13.570062288 +0000 UTC m=+1.839394578,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.747368 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299ae96a696a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.571610986 +0000 UTC m=+1.840943276,LastTimestamp:2026-03-16 00:06:13.571610986 +0000 UTC m=+1.840943276,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.751194 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299ae96f2cd9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.571923161 +0000 UTC m=+1.841255471,LastTimestamp:2026-03-16 00:06:13.571923161 +0000 UTC m=+1.841255471,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.755183 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d299ae978a288 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.572543112 +0000 UTC m=+1.841875412,LastTimestamp:2026-03-16 00:06:13.572543112 +0000 UTC m=+1.841875412,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.758884 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299ae97f475a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.572978522 +0000 UTC m=+1.842310842,LastTimestamp:2026-03-16 00:06:13.572978522 +0000 UTC m=+1.842310842,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.762508 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299aeaeaebf5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.596810229 +0000 UTC m=+1.866142549,LastTimestamp:2026-03-16 00:06:13.596810229 +0000 UTC m=+1.866142549,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.767507 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d299aeaf23ae2 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.597289186 +0000 UTC m=+1.866621516,LastTimestamp:2026-03-16 00:06:13.597289186 +0000 UTC m=+1.866621516,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.772768 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299afc6d7d5f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.890579807 +0000 UTC m=+2.159912137,LastTimestamp:2026-03-16 00:06:13.890579807 +0000 UTC m=+2.159912137,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.783435 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299afcf8e528 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.89971588 +0000 UTC m=+2.169048180,LastTimestamp:2026-03-16 00:06:13.89971588 +0000 UTC m=+2.169048180,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.791219 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299afd0de559 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.901092185 +0000 UTC m=+2.170424515,LastTimestamp:2026-03-16 00:06:13.901092185 +0000 UTC m=+2.170424515,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.796395 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299b0a600abc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.124579516 +0000 UTC m=+2.393911836,LastTimestamp:2026-03-16 00:06:14.124579516 +0000 UTC m=+2.393911836,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.804622 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299b0b3f3455 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.139204693 +0000 UTC m=+2.408537023,LastTimestamp:2026-03-16 00:06:14.139204693 +0000 UTC m=+2.408537023,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.810016 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299b0b5397e5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.140540901 +0000 UTC m=+2.409873231,LastTimestamp:2026-03-16 00:06:14.140540901 +0000 UTC m=+2.409873231,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.814949 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299b19a95eed openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.381043437 +0000 UTC m=+2.650375757,LastTimestamp:2026-03-16 00:06:14.381043437 +0000 UTC m=+2.650375757,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.820561 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299b1a7de59a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.394971546 +0000 UTC m=+2.664303866,LastTimestamp:2026-03-16 00:06:14.394971546 +0000 UTC m=+2.664303866,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.825616 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299b1f5e19c1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.476773825 +0000 UTC m=+2.746106145,LastTimestamp:2026-03-16 00:06:14.476773825 +0000 UTC m=+2.746106145,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.830979 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b2028e0b0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.490063024 +0000 UTC m=+2.759395344,LastTimestamp:2026-03-16 00:06:14.490063024 +0000 UTC m=+2.759395344,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.837143 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299b20c51180 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.500299136 +0000 UTC m=+2.769631466,LastTimestamp:2026-03-16 00:06:14.500299136 +0000 UTC m=+2.769631466,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.844517 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d299b20c681ba openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.500393402 +0000 UTC m=+2.769725722,LastTimestamp:2026-03-16 00:06:14.500393402 +0000 UTC m=+2.769725722,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.849730 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299b2eeb72e8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.737695464 +0000 UTC m=+3.007027764,LastTimestamp:2026-03-16 00:06:14.737695464 +0000 UTC m=+3.007027764,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.854269 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b2f000bee openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.739045358 +0000 UTC m=+3.008377648,LastTimestamp:2026-03-16 00:06:14.739045358 +0000 UTC m=+3.008377648,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.860457 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299b2fa0e89f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.749587615 +0000 UTC m=+3.018919905,LastTimestamp:2026-03-16 00:06:14.749587615 +0000 UTC m=+3.018919905,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.865884 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299b2faeb022 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.750490658 +0000 UTC m=+3.019822948,LastTimestamp:2026-03-16 00:06:14.750490658 +0000 UTC m=+3.019822948,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.870815 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299b2fcf18a2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.752614562 +0000 UTC m=+3.021946852,LastTimestamp:2026-03-16 00:06:14.752614562 +0000 UTC m=+3.021946852,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.875158 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b2fe58021 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.754082849 +0000 UTC m=+3.023415139,LastTimestamp:2026-03-16 00:06:14.754082849 +0000 UTC m=+3.023415139,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.880068 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d299b2fe6e78d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.754174861 +0000 UTC m=+3.023507151,LastTimestamp:2026-03-16 00:06:14.754174861 +0000 UTC m=+3.023507151,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.887708 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b302021b0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.757925296 +0000 UTC m=+3.027257586,LastTimestamp:2026-03-16 00:06:14.757925296 +0000 UTC m=+3.027257586,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.892898 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d299b3158dce8 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.778420456 +0000 UTC m=+3.047752746,LastTimestamp:2026-03-16 00:06:14.778420456 +0000 UTC m=+3.047752746,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.898491 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299b327f8f0d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.797733645 +0000 UTC m=+3.067065925,LastTimestamp:2026-03-16 00:06:14.797733645 +0000 UTC m=+3.067065925,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.903344 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299b3cd2c0e7 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.970958055 +0000 UTC m=+3.240290345,LastTimestamp:2026-03-16 00:06:14.970958055 +0000 UTC m=+3.240290345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.909911 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b3d6be951 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.980995409 +0000 UTC m=+3.250327709,LastTimestamp:2026-03-16 00:06:14.980995409 +0000 UTC m=+3.250327709,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.914445 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299b3d9cd184 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.98420058 +0000 UTC m=+3.253532880,LastTimestamp:2026-03-16 00:06:14.98420058 +0000 UTC m=+3.253532880,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.919387 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299b3dadb11e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.985306398 +0000 UTC m=+3.254638688,LastTimestamp:2026-03-16 00:06:14.985306398 +0000 UTC m=+3.254638688,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.923834 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b3e7ae6b7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.998754999 +0000 UTC m=+3.268087289,LastTimestamp:2026-03-16 00:06:14.998754999 +0000 UTC m=+3.268087289,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.930375 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b3e8998f6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:14.999718134 +0000 UTC m=+3.269050414,LastTimestamp:2026-03-16 00:06:14.999718134 +0000 UTC m=+3.269050414,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.935136 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299b4a7a8f0e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.20005915 +0000 UTC m=+3.469391450,LastTimestamp:2026-03-16 00:06:15.20005915 +0000 UTC m=+3.469391450,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.940297 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b4aacde2c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.203356204 +0000 UTC m=+3.472688494,LastTimestamp:2026-03-16 00:06:15.203356204 +0000 UTC m=+3.472688494,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.944677 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d299b4b63cc12 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.215344658 +0000 UTC m=+3.484676978,LastTimestamp:2026-03-16 00:06:15.215344658 +0000 UTC m=+3.484676978,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.948909 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b4bafe7ce openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.220332494 +0000 UTC m=+3.489664814,LastTimestamp:2026-03-16 00:06:15.220332494 +0000 UTC m=+3.489664814,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.954894 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b4bc5a858 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.22175804 +0000 UTC m=+3.491090370,LastTimestamp:2026-03-16 00:06:15.22175804 +0000 UTC m=+3.491090370,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.961793 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b5929636b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.446397803 +0000 UTC m=+3.715730103,LastTimestamp:2026-03-16 00:06:15.446397803 +0000 UTC m=+3.715730103,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.969168 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b59ecf93e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.459215678 +0000 UTC m=+3.728547968,LastTimestamp:2026-03-16 00:06:15.459215678 +0000 UTC m=+3.728547968,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.973483 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b5a022f39 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.460605753 +0000 UTC m=+3.729938063,LastTimestamp:2026-03-16 00:06:15.460605753 +0000 UTC m=+3.729938063,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.978202 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299b5c9f86fb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.504471803 +0000 UTC m=+3.773804093,LastTimestamp:2026-03-16 00:06:15.504471803 +0000 UTC m=+3.773804093,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.983419 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b676bfb42 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.685643074 +0000 UTC m=+3.954975364,LastTimestamp:2026-03-16 00:06:15.685643074 +0000 UTC m=+3.954975364,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.988135 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299b677f1be8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.686896616 +0000 UTC m=+3.956228926,LastTimestamp:2026-03-16 00:06:15.686896616 +0000 UTC m=+3.956228926,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.992889 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b681f0f5a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.697379162 +0000 UTC m=+3.966711462,LastTimestamp:2026-03-16 00:06:15.697379162 +0000 UTC m=+3.966711462,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:16 crc kubenswrapper[4751]: E0316 00:07:16.999327 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299b68902fbc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.70479302 +0000 UTC m=+3.974125310,LastTimestamp:2026-03-16 00:06:15.70479302 +0000 UTC m=+3.974125310,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.007182 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299b99aadbb6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:16.528624566 +0000 UTC m=+4.797956896,LastTimestamp:2026-03-16 00:06:16.528624566 +0000 UTC m=+4.797956896,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.013589 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299ba97f09a4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:16.794188196 +0000 UTC m=+5.063520526,LastTimestamp:2026-03-16 00:06:16.794188196 +0000 UTC m=+5.063520526,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.018059 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299baa33f31e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:16.806044446 +0000 UTC m=+5.075376766,LastTimestamp:2026-03-16 00:06:16.806044446 +0000 UTC m=+5.075376766,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.025126 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299baa596c0e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:16.808500238 +0000 UTC m=+5.077832568,LastTimestamp:2026-03-16 00:06:16.808500238 +0000 UTC m=+5.077832568,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.032395 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bb9fc884c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.070848076 +0000 UTC m=+5.340180396,LastTimestamp:2026-03-16 00:06:17.070848076 +0000 UTC m=+5.340180396,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.039330 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bbafc76a3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.087620771 +0000 UTC m=+5.356953091,LastTimestamp:2026-03-16 00:06:17.087620771 +0000 UTC m=+5.356953091,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.045929 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bbb1428c5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.089173701 +0000 UTC m=+5.358506001,LastTimestamp:2026-03-16 00:06:17.089173701 +0000 UTC m=+5.358506001,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.052979 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bca2639e0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.342015968 +0000 UTC m=+5.611348288,LastTimestamp:2026-03-16 00:06:17.342015968 +0000 UTC m=+5.611348288,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.059124 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bcb407af1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.360513777 +0000 UTC m=+5.629846107,LastTimestamp:2026-03-16 00:06:17.360513777 +0000 UTC m=+5.629846107,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.064688 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bcb56b160 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.361969504 +0000 UTC m=+5.631301834,LastTimestamp:2026-03-16 00:06:17.361969504 +0000 UTC m=+5.631301834,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.071218 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bdbae8cac openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.636162732 +0000 UTC m=+5.905495062,LastTimestamp:2026-03-16 00:06:17.636162732 +0000 UTC m=+5.905495062,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.075790 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bdc829f26 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.650061094 +0000 UTC m=+5.919393434,LastTimestamp:2026-03-16 00:06:17.650061094 +0000 UTC m=+5.919393434,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.081550 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bdc9bcb60 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.651710816 +0000 UTC m=+5.921043136,LastTimestamp:2026-03-16 00:06:17.651710816 +0000 UTC m=+5.921043136,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.085259 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bec7e9a34 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.91823314 +0000 UTC m=+6.187565470,LastTimestamp:2026-03-16 00:06:17.91823314 +0000 UTC m=+6.187565470,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.091189 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d299bed876b99 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:17.935588249 +0000 UTC m=+6.204920579,LastTimestamp:2026-03-16 00:06:17.935588249 +0000 UTC m=+6.204920579,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.100916 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 16 00:07:17 crc kubenswrapper[4751]: &Event{ObjectMeta:{kube-controller-manager-crc.189d299d0f9cbd28 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 16 00:07:17 crc kubenswrapper[4751]: body: Mar 16 00:07:17 crc kubenswrapper[4751]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:22.802378024 +0000 UTC m=+11.071710354,LastTimestamp:2026-03-16 00:06:22.802378024 +0000 UTC m=+11.071710354,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 16 00:07:17 crc kubenswrapper[4751]: > Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.107611 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299d0f9e6c57 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:22.802488407 +0000 UTC m=+11.071820737,LastTimestamp:2026-03-16 00:06:22.802488407 +0000 UTC m=+11.071820737,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.113741 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 16 00:07:17 crc kubenswrapper[4751]: &Event{ObjectMeta:{kube-apiserver-crc.189d299dee43e608 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 16 00:07:17 crc kubenswrapper[4751]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 16 00:07:17 crc kubenswrapper[4751]: Mar 16 00:07:17 crc kubenswrapper[4751]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:26.537874952 +0000 UTC m=+14.807207282,LastTimestamp:2026-03-16 00:06:26.537874952 +0000 UTC m=+14.807207282,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 16 00:07:17 crc kubenswrapper[4751]: > Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.119987 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299dee44f229 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:26.537943593 +0000 UTC m=+14.807275913,LastTimestamp:2026-03-16 00:06:26.537943593 +0000 UTC m=+14.807275913,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.124632 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d299dee43e608\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 16 00:07:17 crc kubenswrapper[4751]: &Event{ObjectMeta:{kube-apiserver-crc.189d299dee43e608 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 16 00:07:17 crc kubenswrapper[4751]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 16 00:07:17 crc kubenswrapper[4751]: Mar 16 00:07:17 crc kubenswrapper[4751]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:26.537874952 +0000 UTC m=+14.807207282,LastTimestamp:2026-03-16 00:06:26.543134835 +0000 UTC m=+14.812467155,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 16 00:07:17 crc kubenswrapper[4751]: > Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.129343 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d299dee44f229\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299dee44f229 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:26.537943593 +0000 UTC m=+14.807275913,LastTimestamp:2026-03-16 00:06:26.543205487 +0000 UTC m=+14.812537817,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.134531 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d299b5a022f39\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b5a022f39 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.460605753 +0000 UTC m=+3.729938063,LastTimestamp:2026-03-16 00:06:27.584585146 +0000 UTC m=+15.853917446,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.141570 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d299b676bfb42\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b676bfb42 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.685643074 +0000 UTC m=+3.954975364,LastTimestamp:2026-03-16 00:06:27.806848708 +0000 UTC m=+16.076181008,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.147935 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d299b681f0f5a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d299b681f0f5a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:15.697379162 +0000 UTC m=+3.966711462,LastTimestamp:2026-03-16 00:06:27.823408158 +0000 UTC m=+16.092740468,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.153451 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 16 00:07:17 crc kubenswrapper[4751]: &Event{ObjectMeta:{kube-controller-manager-crc.189d299f63ab5ca3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 16 00:07:17 crc kubenswrapper[4751]: body: Mar 16 00:07:17 crc kubenswrapper[4751]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:32.802557091 +0000 UTC m=+21.071889411,LastTimestamp:2026-03-16 00:06:32.802557091 +0000 UTC m=+21.071889411,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 16 00:07:17 crc kubenswrapper[4751]: > Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.157862 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299f63acb4d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:32.802645203 +0000 UTC m=+21.071977533,LastTimestamp:2026-03-16 00:06:32.802645203 +0000 UTC m=+21.071977533,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.164156 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d299f63ab5ca3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 16 00:07:17 crc kubenswrapper[4751]: &Event{ObjectMeta:{kube-controller-manager-crc.189d299f63ab5ca3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 16 00:07:17 crc kubenswrapper[4751]: body: Mar 16 00:07:17 crc kubenswrapper[4751]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:32.802557091 +0000 UTC m=+21.071889411,LastTimestamp:2026-03-16 00:06:42.80248475 +0000 UTC m=+31.071817080,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 16 00:07:17 crc kubenswrapper[4751]: > Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.169331 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d299f63acb4d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299f63acb4d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:32.802645203 +0000 UTC m=+21.071977533,LastTimestamp:2026-03-16 00:06:42.802581063 +0000 UTC m=+31.071913393,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.174072 4751 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d29a1b7e45edb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:42.805513947 +0000 UTC m=+31.074846267,LastTimestamp:2026-03-16 00:06:42.805513947 +0000 UTC m=+31.074846267,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.178563 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d299ae97f475a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299ae97f475a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.572978522 +0000 UTC m=+1.842310842,LastTimestamp:2026-03-16 00:06:42.935774688 +0000 UTC m=+31.205107028,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.182541 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d299afc6d7d5f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299afc6d7d5f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.890579807 +0000 UTC m=+2.159912137,LastTimestamp:2026-03-16 00:06:43.204618111 +0000 UTC m=+31.473950441,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.187501 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d299afcf8e528\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299afcf8e528 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:13.89971588 +0000 UTC m=+2.169048180,LastTimestamp:2026-03-16 00:06:43.218931294 +0000 UTC m=+31.488263624,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.193660 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d299f63ab5ca3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 16 00:07:17 crc kubenswrapper[4751]: &Event{ObjectMeta:{kube-controller-manager-crc.189d299f63ab5ca3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 16 00:07:17 crc kubenswrapper[4751]: body: Mar 16 00:07:17 crc kubenswrapper[4751]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:32.802557091 +0000 UTC m=+21.071889411,LastTimestamp:2026-03-16 00:06:52.802192555 +0000 UTC m=+41.071524875,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 16 00:07:17 crc kubenswrapper[4751]: > Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.196901 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d299f63acb4d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d299f63acb4d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:32.802645203 +0000 UTC m=+21.071977533,LastTimestamp:2026-03-16 00:06:52.802261237 +0000 UTC m=+41.071593567,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:07:17 crc kubenswrapper[4751]: E0316 00:07:17.201551 4751 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d299f63ab5ca3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 16 00:07:17 crc kubenswrapper[4751]: &Event{ObjectMeta:{kube-controller-manager-crc.189d299f63ab5ca3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 16 00:07:17 crc kubenswrapper[4751]: body: Mar 16 00:07:17 crc kubenswrapper[4751]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:06:32.802557091 +0000 UTC m=+21.071889411,LastTimestamp:2026-03-16 00:07:02.80266932 +0000 UTC m=+51.072001650,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 16 00:07:17 crc kubenswrapper[4751]: > Mar 16 00:07:17 crc kubenswrapper[4751]: I0316 00:07:17.373002 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:18 crc kubenswrapper[4751]: I0316 00:07:18.371837 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:19 crc kubenswrapper[4751]: I0316 00:07:19.371492 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:19 crc kubenswrapper[4751]: I0316 00:07:19.801780 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:07:19 crc kubenswrapper[4751]: I0316 00:07:19.802161 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:19 crc kubenswrapper[4751]: I0316 00:07:19.803759 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:19 crc kubenswrapper[4751]: I0316 00:07:19.803831 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:19 crc kubenswrapper[4751]: I0316 00:07:19.803851 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.371543 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.454334 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.456297 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.456386 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.456410 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.457621 4751 scope.go:117] "RemoveContainer" containerID="c7331ac777f3e8c653bab9d0bae9dc6e67a6c2d68ba440c06eb083a82867ca3a" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.793904 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.796155 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd"} Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.796351 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.797638 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.797675 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:20 crc kubenswrapper[4751]: I0316 00:07:20.797685 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:21 crc kubenswrapper[4751]: I0316 00:07:21.373200 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:21 crc kubenswrapper[4751]: W0316 00:07:21.639196 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 16 00:07:21 crc kubenswrapper[4751]: E0316 00:07:21.639253 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 16 00:07:21 crc kubenswrapper[4751]: E0316 00:07:21.982029 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 16 00:07:21 crc kubenswrapper[4751]: I0316 00:07:21.997307 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:21 crc kubenswrapper[4751]: I0316 00:07:21.999048 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:21 crc kubenswrapper[4751]: I0316 00:07:21.999142 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:21 crc kubenswrapper[4751]: I0316 00:07:21.999169 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:21 crc kubenswrapper[4751]: I0316 00:07:21.999209 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:07:22 crc kubenswrapper[4751]: E0316 00:07:22.005379 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.369573 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:22 crc kubenswrapper[4751]: E0316 00:07:22.510383 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.801767 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.801856 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.804478 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.805036 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.806746 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd" exitCode=255 Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.806786 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd"} Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.806857 4751 scope.go:117] "RemoveContainer" containerID="c7331ac777f3e8c653bab9d0bae9dc6e67a6c2d68ba440c06eb083a82867ca3a" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.806957 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.807756 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.807789 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.807799 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:22 crc kubenswrapper[4751]: I0316 00:07:22.808322 4751 scope.go:117] "RemoveContainer" containerID="b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd" Mar 16 00:07:22 crc kubenswrapper[4751]: E0316 00:07:22.808492 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:07:23 crc kubenswrapper[4751]: I0316 00:07:23.371742 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:23 crc kubenswrapper[4751]: I0316 00:07:23.469177 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:07:23 crc kubenswrapper[4751]: I0316 00:07:23.469382 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:23 crc kubenswrapper[4751]: I0316 00:07:23.470887 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:23 crc kubenswrapper[4751]: I0316 00:07:23.470974 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:23 crc kubenswrapper[4751]: I0316 00:07:23.470998 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:23 crc kubenswrapper[4751]: I0316 00:07:23.812317 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.370590 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.502250 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.502485 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.504363 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.504429 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.504453 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.505467 4751 scope.go:117] "RemoveContainer" containerID="b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd" Mar 16 00:07:24 crc kubenswrapper[4751]: E0316 00:07:24.505872 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.582352 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.600700 4751 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.827018 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.827278 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.828599 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.828653 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.828736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:24 crc kubenswrapper[4751]: I0316 00:07:24.829551 4751 scope.go:117] "RemoveContainer" containerID="b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd" Mar 16 00:07:24 crc kubenswrapper[4751]: E0316 00:07:24.829839 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:07:25 crc kubenswrapper[4751]: I0316 00:07:25.373206 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:26 crc kubenswrapper[4751]: I0316 00:07:26.371195 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:27 crc kubenswrapper[4751]: I0316 00:07:27.370652 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 16 00:07:27 crc kubenswrapper[4751]: I0316 00:07:27.684377 4751 csr.go:261] certificate signing request csr-2c54m is approved, waiting to be issued Mar 16 00:07:27 crc kubenswrapper[4751]: I0316 00:07:27.695047 4751 csr.go:257] certificate signing request csr-2c54m is issued Mar 16 00:07:27 crc kubenswrapper[4751]: I0316 00:07:27.743497 4751 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 16 00:07:28 crc kubenswrapper[4751]: I0316 00:07:28.207284 4751 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 16 00:07:28 crc kubenswrapper[4751]: I0316 00:07:28.696062 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-05 03:40:49.231944882 +0000 UTC Mar 16 00:07:28 crc kubenswrapper[4751]: I0316 00:07:28.696152 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6339h33m20.535798094s for next certificate rotation Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.005870 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.007450 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.007509 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.007529 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.007674 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.017822 4751 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.018146 4751 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.018181 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.023393 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.023442 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.023461 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.023487 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.023506 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:29Z","lastTransitionTime":"2026-03-16T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.043360 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.054878 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.054950 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.054974 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.055007 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.055030 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:29Z","lastTransitionTime":"2026-03-16T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.074782 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.086632 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.086693 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.086711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.086736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.086755 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:29Z","lastTransitionTime":"2026-03-16T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.104158 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.114954 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.115041 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.115073 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.115099 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.115150 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:29Z","lastTransitionTime":"2026-03-16T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.131583 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.131808 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.131854 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.231975 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.333241 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.433589 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.534703 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.635575 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.736634 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.810674 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.811001 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.812742 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.812813 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.812834 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.817346 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.836751 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.837326 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.839045 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.839138 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:29 crc kubenswrapper[4751]: I0316 00:07:29.839158 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:29 crc kubenswrapper[4751]: E0316 00:07:29.937416 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.038335 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.139469 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.240660 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.341027 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.441711 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.542388 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.643400 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.743683 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.844678 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:30 crc kubenswrapper[4751]: E0316 00:07:30.945201 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.046233 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.147234 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.247364 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.348453 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.449539 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.550709 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.651362 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.752220 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.852404 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:31 crc kubenswrapper[4751]: E0316 00:07:31.953464 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.053723 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.153974 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.254742 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.355727 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.456873 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.511311 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.557299 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.658355 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.758754 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.858992 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:32 crc kubenswrapper[4751]: E0316 00:07:32.960152 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.061336 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.162053 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.262563 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.362813 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.463617 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.564353 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.665502 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.766155 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.867298 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:33 crc kubenswrapper[4751]: I0316 00:07:33.964372 4751 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 16 00:07:33 crc kubenswrapper[4751]: E0316 00:07:33.968170 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.068917 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.169601 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.270685 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.371696 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: I0316 00:07:34.453998 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:34 crc kubenswrapper[4751]: I0316 00:07:34.456081 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:34 crc kubenswrapper[4751]: I0316 00:07:34.456200 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:34 crc kubenswrapper[4751]: I0316 00:07:34.456231 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.471873 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.572677 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.673779 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.774285 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.875361 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:34 crc kubenswrapper[4751]: E0316 00:07:34.975476 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.076645 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.176983 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.278034 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.378229 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.478927 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.579041 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.679894 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.781084 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.881958 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:35 crc kubenswrapper[4751]: E0316 00:07:35.983152 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.084034 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.184757 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.285308 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.386486 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.487299 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.587696 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.687897 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.789004 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.889508 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:36 crc kubenswrapper[4751]: E0316 00:07:36.990537 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.091511 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.192210 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.292808 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.393455 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: I0316 00:07:37.454088 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:37 crc kubenswrapper[4751]: I0316 00:07:37.456054 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:37 crc kubenswrapper[4751]: I0316 00:07:37.456161 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:37 crc kubenswrapper[4751]: I0316 00:07:37.456193 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:37 crc kubenswrapper[4751]: I0316 00:07:37.457405 4751 scope.go:117] "RemoveContainer" containerID="b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.457808 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.494237 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.595014 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.696159 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.797354 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.898397 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:37 crc kubenswrapper[4751]: E0316 00:07:37.999366 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:38 crc kubenswrapper[4751]: E0316 00:07:38.100185 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:38 crc kubenswrapper[4751]: E0316 00:07:38.201063 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:38 crc kubenswrapper[4751]: E0316 00:07:38.301449 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:38 crc kubenswrapper[4751]: E0316 00:07:38.401914 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:38 crc kubenswrapper[4751]: E0316 00:07:38.503218 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:38 crc kubenswrapper[4751]: E0316 00:07:38.604333 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:38 crc kubenswrapper[4751]: E0316 00:07:38.705303 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:38 crc kubenswrapper[4751]: E0316 00:07:38.806260 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:38 crc kubenswrapper[4751]: E0316 00:07:38.907307 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.008437 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.108962 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.209067 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.310068 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.335399 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.342087 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.342182 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.342203 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.342231 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.342251 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:39Z","lastTransitionTime":"2026-03-16T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.359378 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.365373 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.365452 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.365473 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.365507 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.365532 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:39Z","lastTransitionTime":"2026-03-16T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.382838 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.387992 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.388047 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.388068 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.388097 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.388142 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:39Z","lastTransitionTime":"2026-03-16T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.402958 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.407938 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.408017 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.408038 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.408060 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:39 crc kubenswrapper[4751]: I0316 00:07:39.408130 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:39Z","lastTransitionTime":"2026-03-16T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.423781 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.423998 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.424055 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.524737 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.625564 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.726401 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.827572 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:39 crc kubenswrapper[4751]: E0316 00:07:39.928591 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.029176 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.130239 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.231141 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.332018 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.432238 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: I0316 00:07:40.514250 4751 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.532742 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.633182 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.733301 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.834291 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:40 crc kubenswrapper[4751]: E0316 00:07:40.935178 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.036120 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.137169 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.238250 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.339402 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.440590 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.540948 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.641661 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.742130 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.842991 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:41 crc kubenswrapper[4751]: E0316 00:07:41.944179 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.044321 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.145232 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.246317 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.347146 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.447292 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.511890 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.547527 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.648473 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.749558 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.850470 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:42 crc kubenswrapper[4751]: E0316 00:07:42.950744 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: I0316 00:07:43.050215 4751 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.051443 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.152059 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.253302 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.354358 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.455460 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.555734 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.656161 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.757145 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.858233 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:43 crc kubenswrapper[4751]: E0316 00:07:43.958834 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.059767 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.160653 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.261849 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.362190 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.462805 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.563760 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.664397 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.765320 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.865837 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:44 crc kubenswrapper[4751]: E0316 00:07:44.966893 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.067444 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.168398 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.269209 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.370049 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.470936 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.571381 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.672541 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.772753 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.873827 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:45 crc kubenswrapper[4751]: E0316 00:07:45.974725 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.075514 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.175616 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.276402 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.377502 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: I0316 00:07:46.454314 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:46 crc kubenswrapper[4751]: I0316 00:07:46.455887 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:46 crc kubenswrapper[4751]: I0316 00:07:46.456067 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:46 crc kubenswrapper[4751]: I0316 00:07:46.456223 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.477745 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.578865 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.679965 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.780529 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.880631 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:46 crc kubenswrapper[4751]: E0316 00:07:46.981707 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.082552 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.183494 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.283929 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.384882 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.485564 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.586378 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.686806 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.788062 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.889603 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:47 crc kubenswrapper[4751]: E0316 00:07:47.989745 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.090411 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.191157 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.291620 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.392768 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.493265 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.593611 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.694667 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.795064 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.895408 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:48 crc kubenswrapper[4751]: E0316 00:07:48.995559 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.096501 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.197692 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.298807 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.399160 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.499611 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.595472 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.601387 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.601462 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.601491 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.601519 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.601537 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:49Z","lastTransitionTime":"2026-03-16T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.617744 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.622655 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.622704 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.622723 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.622748 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.622766 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:49Z","lastTransitionTime":"2026-03-16T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.638666 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.643412 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.643485 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.643511 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.643545 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.643563 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:49Z","lastTransitionTime":"2026-03-16T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.658599 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.663148 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.663220 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.663243 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.663282 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:49 crc kubenswrapper[4751]: I0316 00:07:49.663304 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:49Z","lastTransitionTime":"2026-03-16T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.674699 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f81d5235-5710-4ce7-825a-012028fcd5a2\\\",\\\"systemUUID\\\":\\\"7dd49885-293a-444a-a9e7-56b3b0908e03\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.674949 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.675002 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.775868 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.876086 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:49 crc kubenswrapper[4751]: E0316 00:07:49.976495 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.076646 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.177689 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.278881 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.379510 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.479843 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.580888 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.682049 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.782863 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.883909 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:50 crc kubenswrapper[4751]: E0316 00:07:50.985016 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.085438 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.185639 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.286038 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.386914 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: I0316 00:07:51.454408 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 16 00:07:51 crc kubenswrapper[4751]: I0316 00:07:51.456306 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:51 crc kubenswrapper[4751]: I0316 00:07:51.456422 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:51 crc kubenswrapper[4751]: I0316 00:07:51.456452 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:51 crc kubenswrapper[4751]: I0316 00:07:51.457924 4751 scope.go:117] "RemoveContainer" containerID="b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.458348 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.487715 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.588753 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.689635 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.790722 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.891001 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:51 crc kubenswrapper[4751]: E0316 00:07:51.991692 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.092215 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.192464 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.292931 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.393612 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.494837 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.512400 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.595534 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.695980 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.796361 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.896805 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:52 crc kubenswrapper[4751]: E0316 00:07:52.997640 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:53 crc kubenswrapper[4751]: E0316 00:07:53.098523 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:53 crc kubenswrapper[4751]: E0316 00:07:53.199154 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:53 crc kubenswrapper[4751]: E0316 00:07:53.299576 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:53 crc kubenswrapper[4751]: E0316 00:07:53.400648 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:53 crc kubenswrapper[4751]: E0316 00:07:53.501751 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:53 crc kubenswrapper[4751]: E0316 00:07:53.602181 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:53 crc kubenswrapper[4751]: E0316 00:07:53.702566 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:53 crc kubenswrapper[4751]: E0316 00:07:53.803210 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 16 00:07:53 crc kubenswrapper[4751]: I0316 00:07:53.889169 4751 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 16 00:07:53 crc kubenswrapper[4751]: I0316 00:07:53.906223 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:53 crc kubenswrapper[4751]: I0316 00:07:53.906315 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:53 crc kubenswrapper[4751]: I0316 00:07:53.906342 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:53 crc kubenswrapper[4751]: I0316 00:07:53.906375 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:53 crc kubenswrapper[4751]: I0316 00:07:53.906396 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:53Z","lastTransitionTime":"2026-03-16T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.009984 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.010101 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.010161 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.010187 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.010207 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.113566 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.113635 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.113655 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.113726 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.113753 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.217412 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.217486 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.217504 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.217534 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.217554 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.320965 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.321046 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.321070 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.321134 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.321159 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.404535 4751 apiserver.go:52] "Watching apiserver" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.410953 4751 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.411290 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.411812 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.411988 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.412278 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.412392 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.412452 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.412517 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.412793 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.412839 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.412945 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.418494 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.418821 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.420312 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.420518 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.420543 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.420732 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.420783 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.421020 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.420975 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.425818 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.425890 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.425907 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.425932 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.425949 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.458476 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.479860 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.483310 4751 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.493526 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.508782 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.519854 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.519912 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.519945 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.519981 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520013 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520059 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520142 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520192 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520240 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520282 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520328 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520372 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520418 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520461 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520506 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520551 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520600 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520652 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520696 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520819 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.520740 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521036 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521099 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521188 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521238 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521287 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521333 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521398 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521452 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521447 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521606 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521652 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521784 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.521959 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522011 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522063 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522146 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522196 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522246 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522272 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522293 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522428 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522444 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522619 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522680 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522735 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522781 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522827 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522871 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522919 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.522965 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523016 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523059 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523140 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523153 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523191 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523155 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523238 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523281 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523329 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523375 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523418 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523468 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523513 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523525 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523547 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523566 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523616 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523664 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523710 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523749 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523815 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523816 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523866 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523915 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.523961 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524013 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524060 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524159 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524214 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524268 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524315 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524366 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524413 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524468 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524515 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524567 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524598 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524619 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524728 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524780 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524827 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524867 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525188 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525397 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525539 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525580 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525620 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525661 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525697 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525743 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525780 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525867 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525940 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525979 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526016 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526054 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526095 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526161 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526201 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526240 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526277 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526312 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526352 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526391 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526427 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526471 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524633 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524689 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524732 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524744 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524755 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.524894 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525172 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525407 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.525617 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526043 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526054 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526312 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526383 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.526462 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.526575 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:07:55.026536634 +0000 UTC m=+103.295868984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527008 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527070 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527156 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527197 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527235 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527269 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527307 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527343 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527383 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527419 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527534 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527590 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527633 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527675 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527714 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527750 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527792 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527834 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527889 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527948 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527990 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528024 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528059 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528095 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528207 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528263 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528311 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528358 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528406 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528467 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528522 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528579 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528624 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528663 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528706 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528744 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528782 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528816 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528851 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528886 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528922 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528964 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529005 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529062 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529134 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529173 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529214 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529254 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529294 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529332 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529369 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529405 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529443 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529489 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529526 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529564 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529612 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529648 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529704 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529745 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527529 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.527644 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528057 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528144 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528139 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528078 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528474 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528606 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.528792 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.530623 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.532748 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.533313 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.533423 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.533814 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.533847 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.534075 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.534425 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.534471 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.534633 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.534665 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.535607 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.535824 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.535864 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.536505 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.536930 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.536936 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.536938 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.537552 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.538015 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.538000 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.537559 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.538793 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.539070 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.539155 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.539794 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.540485 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.540523 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.540541 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.540567 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.540586 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.540646 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.541267 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.541278 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.543134 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.543438 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.543477 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.543992 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.544191 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.544364 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.544655 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.544707 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.545273 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.545893 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.546249 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.547599 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.547718 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.547931 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.548227 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.548260 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.548652 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.549047 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.549062 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.549847 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550343 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.529783 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550453 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550134 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550505 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550559 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550587 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550636 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550668 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550674 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550746 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550770 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550795 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550876 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550895 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550914 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550937 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550956 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550958 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550976 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.550995 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551016 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551033 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551050 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551066 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551083 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551115 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551133 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551152 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551169 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551187 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551203 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551223 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551241 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551257 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551274 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551324 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551335 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551352 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551536 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551595 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551622 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551647 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551882 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551944 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.551986 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552044 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552093 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552164 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552209 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552006 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552251 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552663 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552264 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552319 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552343 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552217 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552382 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552650 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552830 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.552855 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.552896 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.553031 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.553355 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.553440 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.553447 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:55.053421409 +0000 UTC m=+103.322753729 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.554647 4751 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.558331 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.558389 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.558700 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.565385 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.565796 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.566312 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.566997 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.568054 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.569908 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.570035 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.571478 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.571619 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.571659 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.571799 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.571989 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.572010 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.572482 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.572498 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.572852 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.572930 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.573261 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.573350 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.573473 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:55.073444839 +0000 UTC m=+103.342777169 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.573466 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.573794 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.573842 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.574015 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.574050 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.574077 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.574170 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.574581 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575041 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.575083 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:55.074552152 +0000 UTC m=+103.343884472 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575247 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575282 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575312 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575342 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575452 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575478 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575502 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575770 4751 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575921 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.575984 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576001 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576092 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576188 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576236 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576516 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576783 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576849 4751 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576902 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576933 4751 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576961 4751 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.576993 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.577036 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.577048 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.577071 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.577146 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.577184 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.577178 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.577229 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.577426 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.577456 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.577446 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.577477 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.577489 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.578046 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.578087 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.578483 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.578690 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.578751 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.578922 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.578545 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.579494 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.579632 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.579509 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.580283 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.580342 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.581735 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.581784 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582218 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582260 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582272 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582287 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: E0316 00:07:54.582340 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:55.082307464 +0000 UTC m=+103.351639794 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582388 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582419 4751 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582446 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582471 4751 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582451 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582542 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582601 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.582911 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.583097 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.583163 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.583176 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.583205 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.583246 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.583251 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.584040 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.584406 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.585670 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.584782 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.585817 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.587333 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.583296 4751 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.587951 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.588148 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.588337 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.588480 4751 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.588612 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.588748 4751 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.588867 4751 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.588997 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.589178 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.589329 4751 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.589461 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.589666 4751 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.589832 4751 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.589569 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.589687 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.590355 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.590555 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.590713 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.590841 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.590968 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.591138 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.591295 4751 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.591424 4751 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.591551 4751 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.591669 4751 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.591783 4751 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.591896 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.592035 4751 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.592247 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.592386 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.592504 4751 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.592632 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.592773 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.592894 4751 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.593027 4751 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.593236 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.593376 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.593493 4751 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.593625 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.594447 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.594615 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.594907 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.595059 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.595273 4751 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.592040 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.593557 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.595553 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.593611 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.593683 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.593735 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.594437 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.594487 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.594498 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.594853 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.594975 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.595200 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.596462 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.596604 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.596741 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.596877 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.596996 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.597241 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.597414 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.597559 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.597719 4751 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.597887 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.598072 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.598240 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.597481 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.598359 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.598831 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.598914 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.599244 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.599240 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.599355 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.599761 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.601255 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.603431 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.603550 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.604035 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.605062 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.605137 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.605387 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.605523 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.605849 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.605860 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.606832 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.607048 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.607051 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.607147 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.607675 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.608379 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.613025 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.630383 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.640824 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.644249 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.644294 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.644405 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.644427 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.644465 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.644482 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699040 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699156 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699256 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699251 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699350 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699280 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699451 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699469 4751 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699483 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699501 4751 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699515 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699527 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699539 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699552 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699564 4751 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699576 4751 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699587 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699599 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699610 4751 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699622 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699634 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699647 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699661 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699674 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699686 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699698 4751 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699711 4751 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699723 4751 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699737 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699749 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699761 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699774 4751 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699785 4751 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699798 4751 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699809 4751 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699821 4751 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699833 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699846 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699858 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699870 4751 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699882 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699894 4751 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699907 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699920 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699933 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699946 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699958 4751 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699972 4751 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699984 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.699997 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700009 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700020 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700032 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700044 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700056 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700067 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700078 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700090 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700133 4751 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700156 4751 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700169 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700181 4751 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700193 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700205 4751 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700217 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700231 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700242 4751 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700254 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700266 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700277 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700289 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700301 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700312 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700325 4751 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700337 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700349 4751 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700360 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700371 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700383 4751 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700396 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700408 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700422 4751 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700498 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700511 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700549 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700562 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700575 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700587 4751 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700601 4751 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700614 4751 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700627 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700639 4751 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700650 4751 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700664 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700674 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700685 4751 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700696 4751 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700708 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700720 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700733 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700745 4751 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700756 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700768 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700779 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700790 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700801 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.700814 4751 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.740821 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.747675 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.747737 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.747754 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.747785 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.747803 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.756987 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.767304 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 16 00:07:54 crc kubenswrapper[4751]: W0316 00:07:54.776820 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-79e31c22180682af1505083ebdebca10c848eb88caae7218f7a5642a018eff3a WatchSource:0}: Error finding container 79e31c22180682af1505083ebdebca10c848eb88caae7218f7a5642a018eff3a: Status 404 returned error can't find the container with id 79e31c22180682af1505083ebdebca10c848eb88caae7218f7a5642a018eff3a Mar 16 00:07:54 crc kubenswrapper[4751]: W0316 00:07:54.790486 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-34409c6650400271204ef9b5f5ff6a624a62c00b21bafd012e5d12a143b1755f WatchSource:0}: Error finding container 34409c6650400271204ef9b5f5ff6a624a62c00b21bafd012e5d12a143b1755f: Status 404 returned error can't find the container with id 34409c6650400271204ef9b5f5ff6a624a62c00b21bafd012e5d12a143b1755f Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.851234 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.851274 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.851286 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.851306 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.851319 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.921855 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"79e31c22180682af1505083ebdebca10c848eb88caae7218f7a5642a018eff3a"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.923563 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7a78d03e9b66fd8dbf48e23c46ba030de79a2ca4b85f0fe6c887de7a58795343"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.924890 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"34409c6650400271204ef9b5f5ff6a624a62c00b21bafd012e5d12a143b1755f"} Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.954956 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.955017 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.955039 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.955071 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:54 crc kubenswrapper[4751]: I0316 00:07:54.955091 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:54Z","lastTransitionTime":"2026-03-16T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.058581 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.058622 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.058632 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.058650 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.058699 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.106002 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.106150 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.106253 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106313 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:07:56.106272591 +0000 UTC m=+104.375604931 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106386 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106396 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106442 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106464 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106484 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:56.106455866 +0000 UTC m=+104.375788206 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106548 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:56.106521008 +0000 UTC m=+104.375853338 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.106396 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106599 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106633 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.106639 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106656 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106719 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:56.106701553 +0000 UTC m=+104.376033883 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106739 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:07:55 crc kubenswrapper[4751]: E0316 00:07:55.106789 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:56.106773716 +0000 UTC m=+104.376106036 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.162071 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.162151 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.162188 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.162220 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.162269 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.266180 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.266315 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.266338 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.266365 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.266384 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.370210 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.370279 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.370297 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.370325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.370383 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.474025 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.474089 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.474138 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.474168 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.474186 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.576880 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.576974 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.576991 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.577022 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.577043 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.680381 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.680445 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.680465 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.680492 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.680511 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.784343 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.784397 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.784409 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.784430 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.784445 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.892637 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.892693 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.892706 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.892729 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.892743 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.928974 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"51c7f06a1f3b1a595d97fdf0536b53ae2cdc9ff1a9131a9a15be52b1e2550ade"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.931766 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3d672a397722075aede53d1e2a1fd926bc44996eb5d44cd190cb2f2ff231fdd9"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.931843 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d34a9d592183402070e98e1bc71779b72f01ca1afeb3ce0c38e494b69efdae8b"} Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.958155 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c7f06a1f3b1a595d97fdf0536b53ae2cdc9ff1a9131a9a15be52b1e2550ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-16T00:07:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:55Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.982741 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:55Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.995842 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.995897 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.995918 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.995998 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:55 crc kubenswrapper[4751]: I0316 00:07:55.996043 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:55Z","lastTransitionTime":"2026-03-16T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.001885 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:55Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.021298 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:56Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.047582 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:56Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.064073 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:56Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.084622 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c7f06a1f3b1a595d97fdf0536b53ae2cdc9ff1a9131a9a15be52b1e2550ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-16T00:07:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:56Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.099061 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.099149 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.099164 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.099189 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.099204 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:56Z","lastTransitionTime":"2026-03-16T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.100301 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:56Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.115885 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:56Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.117183 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.117293 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.117342 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.117383 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.117425 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117522 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117561 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117586 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117579 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117672 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:58.117643548 +0000 UTC m=+106.386975848 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117592 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117709 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:58.11769544 +0000 UTC m=+106.387027740 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117711 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117741 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117761 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117762 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:58.117729881 +0000 UTC m=+106.387062201 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117832 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-16 00:07:58.117809583 +0000 UTC m=+106.387141913 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.117861 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:07:58.117843544 +0000 UTC m=+106.387175874 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.138857 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:56Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.158712 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d672a397722075aede53d1e2a1fd926bc44996eb5d44cd190cb2f2ff231fdd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-16T00:07:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d34a9d592183402070e98e1bc71779b72f01ca1afeb3ce0c38e494b69efdae8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-16T00:07:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:56Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.172601 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-16T00:07:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-16T00:07:56Z is after 2025-08-24T17:21:41Z" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.202041 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.202150 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.202174 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.202201 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.202220 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:56Z","lastTransitionTime":"2026-03-16T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.305472 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.305549 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.305567 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.305595 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.305615 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:56Z","lastTransitionTime":"2026-03-16T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.408961 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.409033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.409055 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.409086 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.409140 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:56Z","lastTransitionTime":"2026-03-16T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.454596 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.454751 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.455029 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.455061 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.455322 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 16 00:07:56 crc kubenswrapper[4751]: E0316 00:07:56.455464 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.462314 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.463150 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.464412 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.465190 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.466358 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.466992 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.467645 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.469025 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.470629 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.472651 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.473690 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.476341 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.477559 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.478704 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.480866 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.482165 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.486992 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.487831 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.489901 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.491198 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.492076 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.493949 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.494889 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.496367 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.497254 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.498601 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.499967 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.501054 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.504510 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.505484 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.506526 4751 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.507899 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.512452 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.513614 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.514494 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.514557 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.514610 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.514628 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.514655 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.514674 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:56Z","lastTransitionTime":"2026-03-16T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.519408 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.521317 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.523634 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.525281 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.527805 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.529287 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.530710 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.532799 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.534083 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.535898 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.537029 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.538870 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.540456 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.542636 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.543680 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.545512 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.546633 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.548289 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.550040 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.616859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.616925 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.616942 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.616970 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.616989 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:56Z","lastTransitionTime":"2026-03-16T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.720159 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.720236 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.720261 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.720298 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.720323 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:56Z","lastTransitionTime":"2026-03-16T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.822501 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.822564 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.822579 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.822604 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.822619 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:56Z","lastTransitionTime":"2026-03-16T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.936312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.936389 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.936413 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.936440 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:56 crc kubenswrapper[4751]: I0316 00:07:56.936461 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:56Z","lastTransitionTime":"2026-03-16T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.042083 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.042161 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.042178 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.042203 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.042220 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.145813 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.145882 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.145900 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.145934 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.145953 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.249638 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.249690 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.249702 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.249721 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.249735 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.352967 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.353014 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.353026 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.353046 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.353060 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.456429 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.456481 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.456493 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.456515 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.456529 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.493548 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-lc9lj"] Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.494141 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lc9lj" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.497030 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.498199 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.498642 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.504612 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-pdtst"] Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.505312 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.507946 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.508922 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.509229 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.509262 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.509416 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-7mrzv"] Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.509816 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.510149 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-6fccz"] Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.510417 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.510447 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.514425 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.514785 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.515163 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.515741 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.515982 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.516228 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.516409 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.536711 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fsd7v"] Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.538798 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.542817 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.542870 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.542815 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.542960 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.543649 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545230 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/91e843e9-b5df-4fa6-9e1a-30f594544ccf-hosts-file\") pod \"node-resolver-lc9lj\" (UID: \"91e843e9-b5df-4fa6-9e1a-30f594544ccf\") " pod="openshift-dns/node-resolver-lc9lj" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545369 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-cni-dir\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545409 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-var-lib-cni-multus\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545451 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th664\" (UniqueName: \"kubernetes.io/projected/91e843e9-b5df-4fa6-9e1a-30f594544ccf-kube-api-access-th664\") pod \"node-resolver-lc9lj\" (UID: \"91e843e9-b5df-4fa6-9e1a-30f594544ccf\") " pod="openshift-dns/node-resolver-lc9lj" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545504 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-system-cni-dir\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545536 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-var-lib-kubelet\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545585 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-hostroot\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545636 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-socket-dir-parent\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545680 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-run-multus-certs\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545725 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-etc-kubernetes\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545760 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-conf-dir\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545793 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-cnibin\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545830 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-run-k8s-cni-cncf-io\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545864 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-run-netns\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545894 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-var-lib-cni-bin\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545927 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-daemon-config\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545957 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vkqm\" (UniqueName: \"kubernetes.io/projected/caf1b5ff-47c0-474b-8429-96f4881a94c7-kube-api-access-9vkqm\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.545991 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/caf1b5ff-47c0-474b-8429-96f4881a94c7-cni-binary-copy\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.546045 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-os-release\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.547307 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.547777 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.561486 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.561576 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.561599 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.561755 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.561802 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647425 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-netns\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647486 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-script-lib\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647517 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8kzt\" (UniqueName: \"kubernetes.io/projected/25429719-d18d-4909-8455-701d09423c36-kube-api-access-v8kzt\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647585 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/caf1b5ff-47c0-474b-8429-96f4881a94c7-cni-binary-copy\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647621 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-bin\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647649 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rkfx\" (UniqueName: \"kubernetes.io/projected/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-kube-api-access-8rkfx\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647681 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-var-lib-cni-multus\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647718 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-kubelet\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647742 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-env-overrides\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647767 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/91e843e9-b5df-4fa6-9e1a-30f594544ccf-hosts-file\") pod \"node-resolver-lc9lj\" (UID: \"91e843e9-b5df-4fa6-9e1a-30f594544ccf\") " pod="openshift-dns/node-resolver-lc9lj" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647790 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-config\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647813 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e828458-13b3-4c40-a29d-849625254883-ovn-node-metrics-cert\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647836 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbzpx\" (UniqueName: \"kubernetes.io/projected/3e828458-13b3-4c40-a29d-849625254883-kube-api-access-fbzpx\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647863 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/25429719-d18d-4909-8455-701d09423c36-proxy-tls\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647889 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th664\" (UniqueName: \"kubernetes.io/projected/91e843e9-b5df-4fa6-9e1a-30f594544ccf-kube-api-access-th664\") pod \"node-resolver-lc9lj\" (UID: \"91e843e9-b5df-4fa6-9e1a-30f594544ccf\") " pod="openshift-dns/node-resolver-lc9lj" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647914 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-system-cni-dir\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647937 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-cnibin\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647963 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/25429719-d18d-4909-8455-701d09423c36-rootfs\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.647989 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-system-cni-dir\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648016 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-hostroot\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648041 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-openvswitch\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648063 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-netd\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648089 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-etc-kubernetes\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648164 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vkqm\" (UniqueName: \"kubernetes.io/projected/caf1b5ff-47c0-474b-8429-96f4881a94c7-kube-api-access-9vkqm\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648191 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648216 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25429719-d18d-4909-8455-701d09423c36-mcd-auth-proxy-config\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648242 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-run-k8s-cni-cncf-io\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648266 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-run-netns\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648292 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-daemon-config\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648316 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-log-socket\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648338 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-slash\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648362 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-systemd-units\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648397 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-os-release\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648421 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-etc-openvswitch\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648443 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-node-log\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648466 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-cni-dir\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648493 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-var-lib-openvswitch\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648514 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648536 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-ovn-kubernetes\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648569 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-var-lib-kubelet\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648592 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648616 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-ovn\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648648 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-cni-binary-copy\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648670 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-socket-dir-parent\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648692 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-run-multus-certs\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648715 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-os-release\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648747 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-cnibin\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648770 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-var-lib-cni-bin\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648793 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-conf-dir\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.648814 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-systemd\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.649748 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/caf1b5ff-47c0-474b-8429-96f4881a94c7-cni-binary-copy\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.649827 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-var-lib-cni-multus\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.649893 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/91e843e9-b5df-4fa6-9e1a-30f594544ccf-hosts-file\") pod \"node-resolver-lc9lj\" (UID: \"91e843e9-b5df-4fa6-9e1a-30f594544ccf\") " pod="openshift-dns/node-resolver-lc9lj" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650362 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-socket-dir-parent\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650403 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-run-multus-certs\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650413 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-system-cni-dir\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650457 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-cnibin\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650421 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-os-release\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650499 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-var-lib-cni-bin\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650514 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-cni-dir\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650537 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-conf-dir\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650576 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-run-k8s-cni-cncf-io\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650525 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-hostroot\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650607 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-var-lib-kubelet\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650624 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-etc-kubernetes\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.650636 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/caf1b5ff-47c0-474b-8429-96f4881a94c7-host-run-netns\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.651293 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/caf1b5ff-47c0-474b-8429-96f4881a94c7-multus-daemon-config\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.665950 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.665987 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.666001 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.666026 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.666042 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.673078 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th664\" (UniqueName: \"kubernetes.io/projected/91e843e9-b5df-4fa6-9e1a-30f594544ccf-kube-api-access-th664\") pod \"node-resolver-lc9lj\" (UID: \"91e843e9-b5df-4fa6-9e1a-30f594544ccf\") " pod="openshift-dns/node-resolver-lc9lj" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.673092 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vkqm\" (UniqueName: \"kubernetes.io/projected/caf1b5ff-47c0-474b-8429-96f4881a94c7-kube-api-access-9vkqm\") pod \"multus-pdtst\" (UID: \"caf1b5ff-47c0-474b-8429-96f4881a94c7\") " pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.749705 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-systemd\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.749767 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8kzt\" (UniqueName: \"kubernetes.io/projected/25429719-d18d-4909-8455-701d09423c36-kube-api-access-v8kzt\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.749791 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-netns\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.749814 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-script-lib\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.749848 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rkfx\" (UniqueName: \"kubernetes.io/projected/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-kube-api-access-8rkfx\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.749874 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-systemd\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.749984 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-netns\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750159 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-bin\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.749880 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-bin\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750424 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-env-overrides\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750452 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-kubelet\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750473 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbzpx\" (UniqueName: \"kubernetes.io/projected/3e828458-13b3-4c40-a29d-849625254883-kube-api-access-fbzpx\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750494 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/25429719-d18d-4909-8455-701d09423c36-proxy-tls\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750515 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-system-cni-dir\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750536 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-cnibin\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750561 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-config\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750582 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e828458-13b3-4c40-a29d-849625254883-ovn-node-metrics-cert\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750607 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/25429719-d18d-4909-8455-701d09423c36-rootfs\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750630 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-netd\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750654 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-openvswitch\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750677 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25429719-d18d-4909-8455-701d09423c36-mcd-auth-proxy-config\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750701 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750713 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-cnibin\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750754 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-slash\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750723 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-slash\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750805 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-log-socket\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750839 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-systemd-units\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750860 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-var-lib-openvswitch\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750880 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-etc-openvswitch\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750901 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-node-log\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750923 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750945 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-ovn-kubernetes\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750975 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751007 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-cni-binary-copy\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751030 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-ovn\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751051 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-os-release\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751065 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-env-overrides\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751142 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-kubelet\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751730 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-system-cni-dir\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.750666 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-script-lib\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751802 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-openvswitch\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751905 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-systemd-units\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751920 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-ovn\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751944 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-ovn-kubernetes\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751975 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/25429719-d18d-4909-8455-701d09423c36-rootfs\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751987 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-os-release\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.751997 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-var-lib-openvswitch\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752024 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-node-log\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752034 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752051 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-etc-openvswitch\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752013 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-log-socket\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752072 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-netd\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752517 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752626 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25429719-d18d-4909-8455-701d09423c36-mcd-auth-proxy-config\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752700 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-config\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752825 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-cni-binary-copy\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.752831 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.756038 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e828458-13b3-4c40-a29d-849625254883-ovn-node-metrics-cert\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.756245 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/25429719-d18d-4909-8455-701d09423c36-proxy-tls\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.769211 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbzpx\" (UniqueName: \"kubernetes.io/projected/3e828458-13b3-4c40-a29d-849625254883-kube-api-access-fbzpx\") pod \"ovnkube-node-fsd7v\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.769835 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.769875 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.769887 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.769910 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.769924 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.771405 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-xknbz"] Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.772150 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.774612 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.774652 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.774872 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.775150 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rkfx\" (UniqueName: \"kubernetes.io/projected/b250de1b-5eaf-46f5-9e68-cc8eb4b33100-kube-api-access-8rkfx\") pod \"multus-additional-cni-plugins-7mrzv\" (UID: \"b250de1b-5eaf-46f5-9e68-cc8eb4b33100\") " pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.776526 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.779336 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8kzt\" (UniqueName: \"kubernetes.io/projected/25429719-d18d-4909-8455-701d09423c36-kube-api-access-v8kzt\") pod \"machine-config-daemon-6fccz\" (UID: \"25429719-d18d-4909-8455-701d09423c36\") " pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.817416 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lc9lj" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.838035 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pdtst" Mar 16 00:07:57 crc kubenswrapper[4751]: W0316 00:07:57.852757 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcaf1b5ff_47c0_474b_8429_96f4881a94c7.slice/crio-37cf87f1368a0a91d5dca17a2a888317ad550e5c0b72a88887c0932a4b4d67d8 WatchSource:0}: Error finding container 37cf87f1368a0a91d5dca17a2a888317ad550e5c0b72a88887c0932a4b4d67d8: Status 404 returned error can't find the container with id 37cf87f1368a0a91d5dca17a2a888317ad550e5c0b72a88887c0932a4b4d67d8 Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.853489 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.853923 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w845\" (UniqueName: \"kubernetes.io/projected/92900f7d-1c24-4e2b-a195-ffdf729a4e5f-kube-api-access-8w845\") pod \"node-ca-xknbz\" (UID: \"92900f7d-1c24-4e2b-a195-ffdf729a4e5f\") " pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.854040 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92900f7d-1c24-4e2b-a195-ffdf729a4e5f-host\") pod \"node-ca-xknbz\" (UID: \"92900f7d-1c24-4e2b-a195-ffdf729a4e5f\") " pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.854251 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/92900f7d-1c24-4e2b-a195-ffdf729a4e5f-serviceca\") pod \"node-ca-xknbz\" (UID: \"92900f7d-1c24-4e2b-a195-ffdf729a4e5f\") " pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.868029 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.874856 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.874905 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.874926 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.874951 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.874970 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: W0316 00:07:57.878237 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25429719_d18d_4909_8455_701d09423c36.slice/crio-0b1570b354d532edf46377172544e8b4efeb3450a9adc8800533ae23b1e794ac WatchSource:0}: Error finding container 0b1570b354d532edf46377172544e8b4efeb3450a9adc8800533ae23b1e794ac: Status 404 returned error can't find the container with id 0b1570b354d532edf46377172544e8b4efeb3450a9adc8800533ae23b1e794ac Mar 16 00:07:57 crc kubenswrapper[4751]: W0316 00:07:57.898220 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb250de1b_5eaf_46f5_9e68_cc8eb4b33100.slice/crio-7387450021ea590966bd4f90dbcceb7cd79df13c54261f6430baf41d27736b26 WatchSource:0}: Error finding container 7387450021ea590966bd4f90dbcceb7cd79df13c54261f6430baf41d27736b26: Status 404 returned error can't find the container with id 7387450021ea590966bd4f90dbcceb7cd79df13c54261f6430baf41d27736b26 Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.910093 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.942980 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" event={"ID":"b250de1b-5eaf-46f5-9e68-cc8eb4b33100","Type":"ContainerStarted","Data":"7387450021ea590966bd4f90dbcceb7cd79df13c54261f6430baf41d27736b26"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.948423 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerStarted","Data":"0b1570b354d532edf46377172544e8b4efeb3450a9adc8800533ae23b1e794ac"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.950022 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pdtst" event={"ID":"caf1b5ff-47c0-474b-8429-96f4881a94c7","Type":"ContainerStarted","Data":"37cf87f1368a0a91d5dca17a2a888317ad550e5c0b72a88887c0932a4b4d67d8"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.954858 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w845\" (UniqueName: \"kubernetes.io/projected/92900f7d-1c24-4e2b-a195-ffdf729a4e5f-kube-api-access-8w845\") pod \"node-ca-xknbz\" (UID: \"92900f7d-1c24-4e2b-a195-ffdf729a4e5f\") " pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.954939 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92900f7d-1c24-4e2b-a195-ffdf729a4e5f-host\") pod \"node-ca-xknbz\" (UID: \"92900f7d-1c24-4e2b-a195-ffdf729a4e5f\") " pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.954974 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/92900f7d-1c24-4e2b-a195-ffdf729a4e5f-serviceca\") pod \"node-ca-xknbz\" (UID: \"92900f7d-1c24-4e2b-a195-ffdf729a4e5f\") " pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.955145 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92900f7d-1c24-4e2b-a195-ffdf729a4e5f-host\") pod \"node-ca-xknbz\" (UID: \"92900f7d-1c24-4e2b-a195-ffdf729a4e5f\") " pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.957052 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/92900f7d-1c24-4e2b-a195-ffdf729a4e5f-serviceca\") pod \"node-ca-xknbz\" (UID: \"92900f7d-1c24-4e2b-a195-ffdf729a4e5f\") " pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.959281 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lc9lj" event={"ID":"91e843e9-b5df-4fa6-9e1a-30f594544ccf","Type":"ContainerStarted","Data":"ad03fb7a486e93b8f5fce6a8877ef6cee5dca70fb75cd873c3d1a9ecb09146ff"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.965268 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2"] Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.966745 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.970571 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.970645 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.982231 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-t2bcd"] Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.982565 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.986448 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.986474 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.986493 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.986507 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:57Z","lastTransitionTime":"2026-03-16T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.986173 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:07:57 crc kubenswrapper[4751]: E0316 00:07:57.987178 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t2bcd" podUID="2eb16148-0952-46b4-8c72-f053bea40c0c" Mar 16 00:07:57 crc kubenswrapper[4751]: I0316 00:07:57.987948 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w845\" (UniqueName: \"kubernetes.io/projected/92900f7d-1c24-4e2b-a195-ffdf729a4e5f-kube-api-access-8w845\") pod \"node-ca-xknbz\" (UID: \"92900f7d-1c24-4e2b-a195-ffdf729a4e5f\") " pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.056291 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.056345 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68jw7\" (UniqueName: \"kubernetes.io/projected/2eb16148-0952-46b4-8c72-f053bea40c0c-kube-api-access-68jw7\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.056398 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/96520fe4-d998-4eac-bdaf-2aed9eead24b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.056421 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/96520fe4-d998-4eac-bdaf-2aed9eead24b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.056443 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pjms\" (UniqueName: \"kubernetes.io/projected/96520fe4-d998-4eac-bdaf-2aed9eead24b-kube-api-access-2pjms\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.056490 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/96520fe4-d998-4eac-bdaf-2aed9eead24b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.087818 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xknbz" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.089134 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.089173 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.089185 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.089202 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.089213 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:58Z","lastTransitionTime":"2026-03-16T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:58 crc kubenswrapper[4751]: W0316 00:07:58.136268 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92900f7d_1c24_4e2b_a195_ffdf729a4e5f.slice/crio-c4b34dede74a4e2d3e6b13ac73fe83578a36b48b1042f0480c1366fec8f4096e WatchSource:0}: Error finding container c4b34dede74a4e2d3e6b13ac73fe83578a36b48b1042f0480c1366fec8f4096e: Status 404 returned error can't find the container with id c4b34dede74a4e2d3e6b13ac73fe83578a36b48b1042f0480c1366fec8f4096e Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.157750 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.157910 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:02.157888549 +0000 UTC m=+110.427220839 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.157964 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.157991 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/96520fe4-d998-4eac-bdaf-2aed9eead24b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.158010 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/96520fe4-d998-4eac-bdaf-2aed9eead24b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.158031 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pjms\" (UniqueName: \"kubernetes.io/projected/96520fe4-d998-4eac-bdaf-2aed9eead24b-kube-api-access-2pjms\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.158058 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.158076 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.158098 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/96520fe4-d998-4eac-bdaf-2aed9eead24b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.158132 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.158152 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.158169 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68jw7\" (UniqueName: \"kubernetes.io/projected/2eb16148-0952-46b4-8c72-f053bea40c0c-kube-api-access-68jw7\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158457 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158532 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:02.158513188 +0000 UTC m=+110.427845518 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158548 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158577 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:02.15856832 +0000 UTC m=+110.427900610 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158652 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158675 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs podName:2eb16148-0952-46b4-8c72-f053bea40c0c nodeName:}" failed. No retries permitted until 2026-03-16 00:07:58.658667403 +0000 UTC m=+106.927999693 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs") pod "network-metrics-daemon-t2bcd" (UID: "2eb16148-0952-46b4-8c72-f053bea40c0c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158726 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158738 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158748 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158768 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:02.158762706 +0000 UTC m=+110.428094996 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158804 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158812 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158818 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.158836 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:02.158829168 +0000 UTC m=+110.428161458 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.158945 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/96520fe4-d998-4eac-bdaf-2aed9eead24b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.159605 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/96520fe4-d998-4eac-bdaf-2aed9eead24b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.164087 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/96520fe4-d998-4eac-bdaf-2aed9eead24b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.177488 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68jw7\" (UniqueName: \"kubernetes.io/projected/2eb16148-0952-46b4-8c72-f053bea40c0c-kube-api-access-68jw7\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.180434 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pjms\" (UniqueName: \"kubernetes.io/projected/96520fe4-d998-4eac-bdaf-2aed9eead24b-kube-api-access-2pjms\") pod \"ovnkube-control-plane-749d76644c-4w7g2\" (UID: \"96520fe4-d998-4eac-bdaf-2aed9eead24b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.192367 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.192426 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.192436 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.192452 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.192464 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:58Z","lastTransitionTime":"2026-03-16T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.287167 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.295920 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.295974 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.295992 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.296020 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.296037 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:58Z","lastTransitionTime":"2026-03-16T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:58 crc kubenswrapper[4751]: W0316 00:07:58.297695 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96520fe4_d998_4eac_bdaf_2aed9eead24b.slice/crio-8ff166eb3df435bc3cdad882f5aea0ae043e806c069bd68a00df7155baffcc4b WatchSource:0}: Error finding container 8ff166eb3df435bc3cdad882f5aea0ae043e806c069bd68a00df7155baffcc4b: Status 404 returned error can't find the container with id 8ff166eb3df435bc3cdad882f5aea0ae043e806c069bd68a00df7155baffcc4b Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.398966 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.399001 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.399011 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.399025 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.399037 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:58Z","lastTransitionTime":"2026-03-16T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.456956 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.457081 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.457452 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.457511 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.457585 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.457630 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.501325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.501374 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.501389 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.501408 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.501418 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:58Z","lastTransitionTime":"2026-03-16T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.603966 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.604010 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.604021 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.604038 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.604064 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:58Z","lastTransitionTime":"2026-03-16T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.662772 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.662993 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: E0316 00:07:58.663166 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs podName:2eb16148-0952-46b4-8c72-f053bea40c0c nodeName:}" failed. No retries permitted until 2026-03-16 00:07:59.663133896 +0000 UTC m=+107.932466226 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs") pod "network-metrics-daemon-t2bcd" (UID: "2eb16148-0952-46b4-8c72-f053bea40c0c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.707378 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.707417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.707429 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.707449 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.707460 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:58Z","lastTransitionTime":"2026-03-16T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.810828 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.810876 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.810888 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.810906 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.810920 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:58Z","lastTransitionTime":"2026-03-16T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.914502 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.914556 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.914568 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.914592 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.914606 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:58Z","lastTransitionTime":"2026-03-16T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.966744 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" event={"ID":"96520fe4-d998-4eac-bdaf-2aed9eead24b","Type":"ContainerStarted","Data":"dae2d80f3703539fde505f850261dd6fba4bc8e24f85425cbcd770931f0262b1"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.966799 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" event={"ID":"96520fe4-d998-4eac-bdaf-2aed9eead24b","Type":"ContainerStarted","Data":"1d77cb0ebee6eaed15667123a0533aa680c7c20bdddc9b9406faf277bd8fa641"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.966814 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" event={"ID":"96520fe4-d998-4eac-bdaf-2aed9eead24b","Type":"ContainerStarted","Data":"8ff166eb3df435bc3cdad882f5aea0ae043e806c069bd68a00df7155baffcc4b"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.970316 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerStarted","Data":"0501eeaaf8a80be7f84fbbd8ca9ea1a002b032be5b408ccbbced16574ead5030"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.970352 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerStarted","Data":"eb29dfbd6c43f45308629c345a9e5bfbd8b4e1c106c34f4c658c32496548540a"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.973242 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pdtst" event={"ID":"caf1b5ff-47c0-474b-8429-96f4881a94c7","Type":"ContainerStarted","Data":"8aa33b1d84d61c1f740aa738253ac05657a70c1515da2181b9fe9441b895e0eb"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.975289 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lc9lj" event={"ID":"91e843e9-b5df-4fa6-9e1a-30f594544ccf","Type":"ContainerStarted","Data":"8c76d2688bc32abecfe77060533855cabf42b93ebec6ec894863494772a8ab53"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.977238 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xknbz" event={"ID":"92900f7d-1c24-4e2b-a195-ffdf729a4e5f","Type":"ContainerStarted","Data":"d28b0d48f1fa2298f876761970a090d49b791dc5b55cba449391a4d3c1480504"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.977275 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xknbz" event={"ID":"92900f7d-1c24-4e2b-a195-ffdf729a4e5f","Type":"ContainerStarted","Data":"c4b34dede74a4e2d3e6b13ac73fe83578a36b48b1042f0480c1366fec8f4096e"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.979256 4751 generic.go:334] "Generic (PLEG): container finished" podID="b250de1b-5eaf-46f5-9e68-cc8eb4b33100" containerID="5ad47dd6f60121b6e7bbac7776a286454d073b5b7b2f5addb5d9a313e3e08e17" exitCode=0 Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.979474 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" event={"ID":"b250de1b-5eaf-46f5-9e68-cc8eb4b33100","Type":"ContainerDied","Data":"5ad47dd6f60121b6e7bbac7776a286454d073b5b7b2f5addb5d9a313e3e08e17"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.981499 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8f16e6ce3ce891176e6c583b9c4619b7b666e9c9e2e0c1c60f906cfae1327460"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.983645 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e828458-13b3-4c40-a29d-849625254883" containerID="1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca" exitCode=0 Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.983687 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.983711 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerStarted","Data":"cec470ab0cd1126c195fed32aad8163cdfba13ca4374503b921df69ffafe7a28"} Mar 16 00:07:58 crc kubenswrapper[4751]: I0316 00:07:58.984823 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4w7g2" podStartSLOduration=31.984807036 podStartE2EDuration="31.984807036s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:07:58.983151746 +0000 UTC m=+107.252484046" watchObservedRunningTime="2026-03-16 00:07:58.984807036 +0000 UTC m=+107.254139336" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.027807 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.027854 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.027870 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.027899 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.027941 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:59Z","lastTransitionTime":"2026-03-16T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.057545 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podStartSLOduration=33.057514123 podStartE2EDuration="33.057514123s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:07:59.038663758 +0000 UTC m=+107.307996058" watchObservedRunningTime="2026-03-16 00:07:59.057514123 +0000 UTC m=+107.326846423" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.081863 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-pdtst" podStartSLOduration=33.0818163 podStartE2EDuration="33.0818163s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:07:59.081291624 +0000 UTC m=+107.350623934" watchObservedRunningTime="2026-03-16 00:07:59.0818163 +0000 UTC m=+107.351148600" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.122444 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-lc9lj" podStartSLOduration=33.122408935 podStartE2EDuration="33.122408935s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:07:59.121184319 +0000 UTC m=+107.390516609" watchObservedRunningTime="2026-03-16 00:07:59.122408935 +0000 UTC m=+107.391741245" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.123794 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-xknbz" podStartSLOduration=33.123779936 podStartE2EDuration="33.123779936s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:07:59.103301723 +0000 UTC m=+107.372634023" watchObservedRunningTime="2026-03-16 00:07:59.123779936 +0000 UTC m=+107.393112246" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.136859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.136898 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.136908 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.136926 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.136938 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:59Z","lastTransitionTime":"2026-03-16T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.242170 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.242239 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.242262 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.242297 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.242318 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:59Z","lastTransitionTime":"2026-03-16T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.345783 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.345840 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.345858 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.345883 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.345906 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:59Z","lastTransitionTime":"2026-03-16T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.449618 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.449672 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.449684 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.449731 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.449744 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:59Z","lastTransitionTime":"2026-03-16T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.453882 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:07:59 crc kubenswrapper[4751]: E0316 00:07:59.454015 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t2bcd" podUID="2eb16148-0952-46b4-8c72-f053bea40c0c" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.556897 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.557371 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.557391 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.557418 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.557436 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:59Z","lastTransitionTime":"2026-03-16T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.660066 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.660118 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.660130 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.660147 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.660157 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:59Z","lastTransitionTime":"2026-03-16T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.674326 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:07:59 crc kubenswrapper[4751]: E0316 00:07:59.674469 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:07:59 crc kubenswrapper[4751]: E0316 00:07:59.674540 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs podName:2eb16148-0952-46b4-8c72-f053bea40c0c nodeName:}" failed. No retries permitted until 2026-03-16 00:08:01.674523754 +0000 UTC m=+109.943856044 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs") pod "network-metrics-daemon-t2bcd" (UID: "2eb16148-0952-46b4-8c72-f053bea40c0c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.684417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.684448 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.684457 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.684472 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.684481 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-16T00:07:59Z","lastTransitionTime":"2026-03-16T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.743848 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6"] Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.744438 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.746811 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.747155 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.747408 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.747939 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.876688 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/592e398a-df7b-4311-985b-0d5c3a36046f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.877493 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/592e398a-df7b-4311-985b-0d5c3a36046f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.877668 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/592e398a-df7b-4311-985b-0d5c3a36046f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.877809 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/592e398a-df7b-4311-985b-0d5c3a36046f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.877915 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/592e398a-df7b-4311-985b-0d5c3a36046f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.979414 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/592e398a-df7b-4311-985b-0d5c3a36046f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.979722 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/592e398a-df7b-4311-985b-0d5c3a36046f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.979969 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/592e398a-df7b-4311-985b-0d5c3a36046f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.980298 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/592e398a-df7b-4311-985b-0d5c3a36046f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.980711 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/592e398a-df7b-4311-985b-0d5c3a36046f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.981308 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/592e398a-df7b-4311-985b-0d5c3a36046f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.980094 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/592e398a-df7b-4311-985b-0d5c3a36046f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.981395 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/592e398a-df7b-4311-985b-0d5c3a36046f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.991436 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/592e398a-df7b-4311-985b-0d5c3a36046f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.994244 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" event={"ID":"b250de1b-5eaf-46f5-9e68-cc8eb4b33100","Type":"ContainerStarted","Data":"aef9d1560809320c806e05d7a116cde139b896efa1793aba1194ccabdd08e492"} Mar 16 00:07:59 crc kubenswrapper[4751]: I0316 00:07:59.998907 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/592e398a-df7b-4311-985b-0d5c3a36046f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-gsch6\" (UID: \"592e398a-df7b-4311-985b-0d5c3a36046f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.003790 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerStarted","Data":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.003861 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerStarted","Data":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.003882 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerStarted","Data":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.003903 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerStarted","Data":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.081581 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" Mar 16 00:08:00 crc kubenswrapper[4751]: W0316 00:08:00.181471 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod592e398a_df7b_4311_985b_0d5c3a36046f.slice/crio-4f469fac8a777958eae4815de0f28836847e5c8efb9e645a1258ec7c1332b78a WatchSource:0}: Error finding container 4f469fac8a777958eae4815de0f28836847e5c8efb9e645a1258ec7c1332b78a: Status 404 returned error can't find the container with id 4f469fac8a777958eae4815de0f28836847e5c8efb9e645a1258ec7c1332b78a Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.445661 4751 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.454786 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.454923 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:00 crc kubenswrapper[4751]: E0316 00:08:00.454994 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 16 00:08:00 crc kubenswrapper[4751]: E0316 00:08:00.455179 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.455277 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:00 crc kubenswrapper[4751]: E0316 00:08:00.455364 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 16 00:08:00 crc kubenswrapper[4751]: I0316 00:08:00.457314 4751 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 16 00:08:01 crc kubenswrapper[4751]: I0316 00:08:01.010560 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" event={"ID":"592e398a-df7b-4311-985b-0d5c3a36046f","Type":"ContainerStarted","Data":"060c26257b3df7eeccc208422d632ae9f655a4b1b27623716e4078df5c88c373"} Mar 16 00:08:01 crc kubenswrapper[4751]: I0316 00:08:01.013718 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" event={"ID":"592e398a-df7b-4311-985b-0d5c3a36046f","Type":"ContainerStarted","Data":"4f469fac8a777958eae4815de0f28836847e5c8efb9e645a1258ec7c1332b78a"} Mar 16 00:08:01 crc kubenswrapper[4751]: I0316 00:08:01.026025 4751 generic.go:334] "Generic (PLEG): container finished" podID="b250de1b-5eaf-46f5-9e68-cc8eb4b33100" containerID="aef9d1560809320c806e05d7a116cde139b896efa1793aba1194ccabdd08e492" exitCode=0 Mar 16 00:08:01 crc kubenswrapper[4751]: I0316 00:08:01.026175 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" event={"ID":"b250de1b-5eaf-46f5-9e68-cc8eb4b33100","Type":"ContainerDied","Data":"aef9d1560809320c806e05d7a116cde139b896efa1793aba1194ccabdd08e492"} Mar 16 00:08:01 crc kubenswrapper[4751]: I0316 00:08:01.033706 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerStarted","Data":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} Mar 16 00:08:01 crc kubenswrapper[4751]: I0316 00:08:01.033756 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerStarted","Data":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} Mar 16 00:08:01 crc kubenswrapper[4751]: I0316 00:08:01.045149 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gsch6" podStartSLOduration=35.045061265 podStartE2EDuration="35.045061265s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:01.044955732 +0000 UTC m=+109.314288062" watchObservedRunningTime="2026-03-16 00:08:01.045061265 +0000 UTC m=+109.314393595" Mar 16 00:08:01 crc kubenswrapper[4751]: I0316 00:08:01.453805 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:01 crc kubenswrapper[4751]: E0316 00:08:01.453992 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t2bcd" podUID="2eb16148-0952-46b4-8c72-f053bea40c0c" Mar 16 00:08:01 crc kubenswrapper[4751]: I0316 00:08:01.706625 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:01 crc kubenswrapper[4751]: E0316 00:08:01.706948 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:08:01 crc kubenswrapper[4751]: E0316 00:08:01.707096 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs podName:2eb16148-0952-46b4-8c72-f053bea40c0c nodeName:}" failed. No retries permitted until 2026-03-16 00:08:05.707059014 +0000 UTC m=+113.976391404 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs") pod "network-metrics-daemon-t2bcd" (UID: "2eb16148-0952-46b4-8c72-f053bea40c0c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.040388 4751 generic.go:334] "Generic (PLEG): container finished" podID="b250de1b-5eaf-46f5-9e68-cc8eb4b33100" containerID="ead603d09db87d45d41b80850a48336e49a32d93b8ff3bdf918525d95c46628b" exitCode=0 Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.040462 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" event={"ID":"b250de1b-5eaf-46f5-9e68-cc8eb4b33100","Type":"ContainerDied","Data":"ead603d09db87d45d41b80850a48336e49a32d93b8ff3bdf918525d95c46628b"} Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.212222 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212379 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:10.212346731 +0000 UTC m=+118.481679031 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.212436 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.212488 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.212568 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.212609 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212611 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212661 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212664 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212678 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212703 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212665 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212736 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212727 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:10.212714512 +0000 UTC m=+118.482046812 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212772 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:10.212761064 +0000 UTC m=+118.482093364 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212786 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:10.212778954 +0000 UTC m=+118.482111254 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.212849 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.213078 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:10.21298455 +0000 UTC m=+118.482316880 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.455486 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.455685 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.456365 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:02 crc kubenswrapper[4751]: I0316 00:08:02.456420 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.456534 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 16 00:08:02 crc kubenswrapper[4751]: E0316 00:08:02.456638 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 16 00:08:03 crc kubenswrapper[4751]: I0316 00:08:03.048026 4751 generic.go:334] "Generic (PLEG): container finished" podID="b250de1b-5eaf-46f5-9e68-cc8eb4b33100" containerID="3834be2d11df912c3c0d19098eb80c4a03f6991af0d4d26b8932bcbfdab73aef" exitCode=0 Mar 16 00:08:03 crc kubenswrapper[4751]: I0316 00:08:03.048133 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" event={"ID":"b250de1b-5eaf-46f5-9e68-cc8eb4b33100","Type":"ContainerDied","Data":"3834be2d11df912c3c0d19098eb80c4a03f6991af0d4d26b8932bcbfdab73aef"} Mar 16 00:08:03 crc kubenswrapper[4751]: I0316 00:08:03.055023 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerStarted","Data":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} Mar 16 00:08:03 crc kubenswrapper[4751]: I0316 00:08:03.453855 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:03 crc kubenswrapper[4751]: E0316 00:08:03.454398 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t2bcd" podUID="2eb16148-0952-46b4-8c72-f053bea40c0c" Mar 16 00:08:04 crc kubenswrapper[4751]: I0316 00:08:04.071025 4751 generic.go:334] "Generic (PLEG): container finished" podID="b250de1b-5eaf-46f5-9e68-cc8eb4b33100" containerID="9ad6a80e9f879159c79d2f002bdad0e59229d389c6627f38ed97a995d791e46d" exitCode=0 Mar 16 00:08:04 crc kubenswrapper[4751]: I0316 00:08:04.071089 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" event={"ID":"b250de1b-5eaf-46f5-9e68-cc8eb4b33100","Type":"ContainerDied","Data":"9ad6a80e9f879159c79d2f002bdad0e59229d389c6627f38ed97a995d791e46d"} Mar 16 00:08:04 crc kubenswrapper[4751]: I0316 00:08:04.454359 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:04 crc kubenswrapper[4751]: I0316 00:08:04.454544 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:04 crc kubenswrapper[4751]: E0316 00:08:04.454670 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 16 00:08:04 crc kubenswrapper[4751]: I0316 00:08:04.454753 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:04 crc kubenswrapper[4751]: E0316 00:08:04.454868 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 16 00:08:04 crc kubenswrapper[4751]: E0316 00:08:04.455045 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.079932 4751 generic.go:334] "Generic (PLEG): container finished" podID="b250de1b-5eaf-46f5-9e68-cc8eb4b33100" containerID="99d76f71cc416b5a4ff5b05cbaa2cd5ff626d6e97eea7754b187921b70dcef13" exitCode=0 Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.080162 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" event={"ID":"b250de1b-5eaf-46f5-9e68-cc8eb4b33100","Type":"ContainerDied","Data":"99d76f71cc416b5a4ff5b05cbaa2cd5ff626d6e97eea7754b187921b70dcef13"} Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.087792 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerStarted","Data":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.088448 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.088563 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.088835 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.161036 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.161202 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.186484 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" podStartSLOduration=39.18645959 podStartE2EDuration="39.18645959s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:05.185615804 +0000 UTC m=+113.454948124" watchObservedRunningTime="2026-03-16 00:08:05.18645959 +0000 UTC m=+113.455791910" Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.454760 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:05 crc kubenswrapper[4751]: E0316 00:08:05.454884 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t2bcd" podUID="2eb16148-0952-46b4-8c72-f053bea40c0c" Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.479148 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.480257 4751 scope.go:117] "RemoveContainer" containerID="b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd" Mar 16 00:08:05 crc kubenswrapper[4751]: I0316 00:08:05.754898 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:05 crc kubenswrapper[4751]: E0316 00:08:05.755080 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:08:05 crc kubenswrapper[4751]: E0316 00:08:05.755166 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs podName:2eb16148-0952-46b4-8c72-f053bea40c0c nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.755147995 +0000 UTC m=+122.024480295 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs") pod "network-metrics-daemon-t2bcd" (UID: "2eb16148-0952-46b4-8c72-f053bea40c0c") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 16 00:08:06 crc kubenswrapper[4751]: I0316 00:08:06.094814 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 16 00:08:06 crc kubenswrapper[4751]: I0316 00:08:06.097567 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807"} Mar 16 00:08:06 crc kubenswrapper[4751]: I0316 00:08:06.098038 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:08:06 crc kubenswrapper[4751]: I0316 00:08:06.104322 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" event={"ID":"b250de1b-5eaf-46f5-9e68-cc8eb4b33100","Type":"ContainerStarted","Data":"04df8348d76cb84fbb82cdc208a64506562275b417b886f105ed4fa29505936f"} Mar 16 00:08:06 crc kubenswrapper[4751]: I0316 00:08:06.133630 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=1.133600875 podStartE2EDuration="1.133600875s" podCreationTimestamp="2026-03-16 00:08:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:06.132305256 +0000 UTC m=+114.401637566" watchObservedRunningTime="2026-03-16 00:08:06.133600875 +0000 UTC m=+114.402933195" Mar 16 00:08:06 crc kubenswrapper[4751]: I0316 00:08:06.169440 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-7mrzv" podStartSLOduration=40.169405747 podStartE2EDuration="40.169405747s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:06.168218922 +0000 UTC m=+114.437551232" watchObservedRunningTime="2026-03-16 00:08:06.169405747 +0000 UTC m=+114.438738077" Mar 16 00:08:06 crc kubenswrapper[4751]: I0316 00:08:06.454495 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:06 crc kubenswrapper[4751]: I0316 00:08:06.454658 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:06 crc kubenswrapper[4751]: E0316 00:08:06.454877 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 16 00:08:06 crc kubenswrapper[4751]: I0316 00:08:06.455548 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:06 crc kubenswrapper[4751]: E0316 00:08:06.455661 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 16 00:08:06 crc kubenswrapper[4751]: E0316 00:08:06.455750 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 16 00:08:07 crc kubenswrapper[4751]: I0316 00:08:07.021056 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-t2bcd"] Mar 16 00:08:07 crc kubenswrapper[4751]: I0316 00:08:07.021502 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:07 crc kubenswrapper[4751]: E0316 00:08:07.021619 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t2bcd" podUID="2eb16148-0952-46b4-8c72-f053bea40c0c" Mar 16 00:08:08 crc kubenswrapper[4751]: I0316 00:08:08.454259 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:08 crc kubenswrapper[4751]: I0316 00:08:08.454332 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:08 crc kubenswrapper[4751]: I0316 00:08:08.454359 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:08 crc kubenswrapper[4751]: E0316 00:08:08.454523 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 16 00:08:08 crc kubenswrapper[4751]: E0316 00:08:08.454658 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 16 00:08:08 crc kubenswrapper[4751]: E0316 00:08:08.454786 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 16 00:08:09 crc kubenswrapper[4751]: I0316 00:08:09.453744 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:09 crc kubenswrapper[4751]: E0316 00:08:09.453969 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t2bcd" podUID="2eb16148-0952-46b4-8c72-f053bea40c0c" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.059488 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.059808 4751 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.123479 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.124391 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.124617 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dklns"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.125217 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.132087 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.132617 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.133246 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vl6bg"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.133920 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.134144 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.134342 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.134442 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.134473 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.137542 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.137561 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.137578 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.137619 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.137744 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.140294 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9mrng"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.140815 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.142422 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zb5z5"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.143063 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.146795 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-88ldb"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.147296 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.148613 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.148904 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.149574 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.151882 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.152116 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.152255 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.152272 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.152408 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.152418 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.152517 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.152649 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.152734 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.152860 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.153041 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.153241 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.153358 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.153427 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.154601 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.155079 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.156660 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.157880 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.167211 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-njjzb"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.168222 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6k5cn"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.169275 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6k5cn" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.169348 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.169475 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.169760 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.171040 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.172279 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.176464 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.177936 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.178283 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.178494 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.178599 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.178678 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.178808 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.178934 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.178961 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.179377 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.179509 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.179554 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.179640 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.180996 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.181695 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.181759 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.187699 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.187891 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.187979 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.190180 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.190303 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.190419 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.190889 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.191454 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.191680 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.193786 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.197571 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.197872 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jszhl"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.197910 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.198336 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.198710 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.198776 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mrm4x"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.199260 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.199556 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.200522 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.200624 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.206602 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/925c7b04-46f9-4ea0-91ce-4f098da81d33-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g5frw\" (UID: \"925c7b04-46f9-4ea0-91ce-4f098da81d33\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.206644 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/925c7b04-46f9-4ea0-91ce-4f098da81d33-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g5frw\" (UID: \"925c7b04-46f9-4ea0-91ce-4f098da81d33\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.206665 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d455ed36-0071-411e-b306-237ca7d6695b-service-ca-bundle\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.206695 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d455ed36-0071-411e-b306-237ca7d6695b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.206731 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d455ed36-0071-411e-b306-237ca7d6695b-serving-cert\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.206761 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t574r\" (UniqueName: \"kubernetes.io/projected/925c7b04-46f9-4ea0-91ce-4f098da81d33-kube-api-access-t574r\") pod \"openshift-apiserver-operator-796bbdcf4f-g5frw\" (UID: \"925c7b04-46f9-4ea0-91ce-4f098da81d33\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.206779 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkn7w\" (UniqueName: \"kubernetes.io/projected/d455ed36-0071-411e-b306-237ca7d6695b-kube-api-access-nkn7w\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.206798 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d455ed36-0071-411e-b306-237ca7d6695b-config\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.206845 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.207310 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.207637 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.207810 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.207858 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.207858 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.208300 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.208312 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.208483 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.208544 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.208664 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.208668 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.208778 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.208829 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.208784 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.209163 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.209259 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.209374 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.210130 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zd5fn"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.210695 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.210754 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.210796 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.210878 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.210988 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.210989 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.211282 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.213025 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.214001 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-bd5b9"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.214361 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.214584 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.215139 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.215272 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.215468 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.219165 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-jpvk6"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.219660 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.219863 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.222598 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.222853 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.223127 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.223310 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.223917 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.224182 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.224289 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.224743 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.224902 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.225032 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29560320-xlv4n"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.225215 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.225314 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.225237 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.225277 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.225709 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.226384 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.226580 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.236478 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.237135 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.240233 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.240641 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.241612 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vl6bg"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.241667 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.244565 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.245473 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.245717 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.248995 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.257972 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.259734 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zb5z5"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.261638 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.268085 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.268792 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sph8v"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.269289 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.269555 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.270827 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.286356 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.286993 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-njjzb"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.287581 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.288141 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.289023 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.289130 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.289266 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.290199 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9mrng"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.290648 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vg885"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.290877 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.291379 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.292002 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dklns"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.292490 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6qwkx"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.293690 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.294059 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.294074 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.294425 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.295511 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.298435 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sfnlp"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.298835 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.300162 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.301056 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.301655 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.303449 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nx86z"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.303895 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.304002 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.305549 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.306058 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.306682 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307193 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307285 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn2sw\" (UniqueName: \"kubernetes.io/projected/2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f-kube-api-access-cn2sw\") pod \"dns-operator-744455d44c-vl6bg\" (UID: \"2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307322 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.307343 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:26.307320007 +0000 UTC m=+134.576652357 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307378 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d455ed36-0071-411e-b306-237ca7d6695b-serving-cert\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307429 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da8ac6ff-99fb-4c50-a706-eec6c8987b04-config\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307457 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-config\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307484 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-audit-dir\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307511 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t574r\" (UniqueName: \"kubernetes.io/projected/925c7b04-46f9-4ea0-91ce-4f098da81d33-kube-api-access-t574r\") pod \"openshift-apiserver-operator-796bbdcf4f-g5frw\" (UID: \"925c7b04-46f9-4ea0-91ce-4f098da81d33\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307533 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7e4947be-bb74-46c6-9a89-9c44dede6c7c-etcd-ca\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307554 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fs9l\" (UniqueName: \"kubernetes.io/projected/5c2eff22-d730-456b-870d-4138d9354577-kube-api-access-8fs9l\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307582 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d455ed36-0071-411e-b306-237ca7d6695b-config\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307606 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-audit-policies\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307627 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4947be-bb74-46c6-9a89-9c44dede6c7c-config\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307647 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7e4947be-bb74-46c6-9a89-9c44dede6c7c-etcd-service-ca\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307667 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c2eff22-d730-456b-870d-4138d9354577-serving-cert\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307690 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da8ac6ff-99fb-4c50-a706-eec6c8987b04-serving-cert\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307713 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8kxm\" (UniqueName: \"kubernetes.io/projected/7e4947be-bb74-46c6-9a89-9c44dede6c7c-kube-api-access-l8kxm\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307752 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f-metrics-tls\") pod \"dns-operator-744455d44c-vl6bg\" (UID: \"2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307775 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k42kp\" (UniqueName: \"kubernetes.io/projected/90685c8c-a90e-4821-9e19-40a677b71265-kube-api-access-k42kp\") pod \"image-pruner-29560320-xlv4n\" (UID: \"90685c8c-a90e-4821-9e19-40a677b71265\") " pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307798 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307818 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-client-ca\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307845 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bzhm\" (UniqueName: \"kubernetes.io/projected/da8ac6ff-99fb-4c50-a706-eec6c8987b04-kube-api-access-8bzhm\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307872 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307901 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307930 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmgbb\" (UniqueName: \"kubernetes.io/projected/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-kube-api-access-fmgbb\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307969 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-encryption-config\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.307998 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkn7w\" (UniqueName: \"kubernetes.io/projected/d455ed36-0071-411e-b306-237ca7d6695b-kube-api-access-nkn7w\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308021 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-serving-cert\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308048 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-etcd-client\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308070 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da8ac6ff-99fb-4c50-a706-eec6c8987b04-trusted-ca\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308093 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e4947be-bb74-46c6-9a89-9c44dede6c7c-serving-cert\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308137 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/925c7b04-46f9-4ea0-91ce-4f098da81d33-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g5frw\" (UID: \"925c7b04-46f9-4ea0-91ce-4f098da81d33\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308160 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/925c7b04-46f9-4ea0-91ce-4f098da81d33-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g5frw\" (UID: \"925c7b04-46f9-4ea0-91ce-4f098da81d33\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308183 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d455ed36-0071-411e-b306-237ca7d6695b-service-ca-bundle\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308209 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308232 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308252 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7e4947be-bb74-46c6-9a89-9c44dede6c7c-etcd-client\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308277 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d455ed36-0071-411e-b306-237ca7d6695b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.308302 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/90685c8c-a90e-4821-9e19-40a677b71265-serviceca\") pod \"image-pruner-29560320-xlv4n\" (UID: \"90685c8c-a90e-4821-9e19-40a677b71265\") " pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.307386 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.308397 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:26.308386319 +0000 UTC m=+134.577718679 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.309493 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.309529 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.309545 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.309605 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:26.309585905 +0000 UTC m=+134.578918255 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.309658 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d455ed36-0071-411e-b306-237ca7d6695b-config\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.310630 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.310777 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:26.31075175 +0000 UTC m=+134.580084090 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.310968 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.311069 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.311173 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:08:10 crc kubenswrapper[4751]: E0316 00:08:10.311348 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:26.311334047 +0000 UTC m=+134.580666397 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.311347 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d455ed36-0071-411e-b306-237ca7d6695b-service-ca-bundle\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.311624 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/925c7b04-46f9-4ea0-91ce-4f098da81d33-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g5frw\" (UID: \"925c7b04-46f9-4ea0-91ce-4f098da81d33\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.312593 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d455ed36-0071-411e-b306-237ca7d6695b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.312645 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.313129 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.313615 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d455ed36-0071-411e-b306-237ca7d6695b-serving-cert\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.313699 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/925c7b04-46f9-4ea0-91ce-4f098da81d33-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g5frw\" (UID: \"925c7b04-46f9-4ea0-91ce-4f098da81d33\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.313802 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-znlfm"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.314824 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.322098 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-nsnh2"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.322980 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-p56l4"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.323384 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.323839 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.324515 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560328-lt56z"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.327055 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.329837 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.329865 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.329955 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mrm4x"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.329931 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560328-lt56z" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.331196 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.334610 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-hz7xk"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.335165 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.336713 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.338518 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.349063 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-69wfh"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.349438 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.352738 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.352794 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.352943 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.356184 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.357396 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jszhl"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.358596 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-88ldb"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.363324 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bd5b9"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.367829 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.368055 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vg885"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.369788 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6qwkx"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.370117 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sph8v"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.371960 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6k5cn"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.372344 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.373294 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.374094 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zd5fn"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.374968 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.376063 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sfnlp"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.376695 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.377523 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560328-lt56z"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.378385 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29560320-xlv4n"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.379367 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.380231 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.381073 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-69wfh"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.381969 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.383265 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nx86z"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.383891 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-znlfm"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.384729 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.385572 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nsnh2"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.386406 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.387234 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6d776"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.387460 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.388387 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6d776"] Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.388618 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6d776" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.407061 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409226 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-audit-policies\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409266 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4947be-bb74-46c6-9a89-9c44dede6c7c-config\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409288 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7e4947be-bb74-46c6-9a89-9c44dede6c7c-etcd-service-ca\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409305 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c2eff22-d730-456b-870d-4138d9354577-serving-cert\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409322 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da8ac6ff-99fb-4c50-a706-eec6c8987b04-serving-cert\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409340 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8kxm\" (UniqueName: \"kubernetes.io/projected/7e4947be-bb74-46c6-9a89-9c44dede6c7c-kube-api-access-l8kxm\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409360 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f-metrics-tls\") pod \"dns-operator-744455d44c-vl6bg\" (UID: \"2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409385 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k42kp\" (UniqueName: \"kubernetes.io/projected/90685c8c-a90e-4821-9e19-40a677b71265-kube-api-access-k42kp\") pod \"image-pruner-29560320-xlv4n\" (UID: \"90685c8c-a90e-4821-9e19-40a677b71265\") " pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409400 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409421 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-client-ca\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409439 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bzhm\" (UniqueName: \"kubernetes.io/projected/da8ac6ff-99fb-4c50-a706-eec6c8987b04-kube-api-access-8bzhm\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409470 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmgbb\" (UniqueName: \"kubernetes.io/projected/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-kube-api-access-fmgbb\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409493 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-encryption-config\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409515 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-serving-cert\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409534 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-etcd-client\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409550 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da8ac6ff-99fb-4c50-a706-eec6c8987b04-trusted-ca\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409564 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e4947be-bb74-46c6-9a89-9c44dede6c7c-serving-cert\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409592 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409610 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7e4947be-bb74-46c6-9a89-9c44dede6c7c-etcd-client\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409627 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/90685c8c-a90e-4821-9e19-40a677b71265-serviceca\") pod \"image-pruner-29560320-xlv4n\" (UID: \"90685c8c-a90e-4821-9e19-40a677b71265\") " pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409644 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn2sw\" (UniqueName: \"kubernetes.io/projected/2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f-kube-api-access-cn2sw\") pod \"dns-operator-744455d44c-vl6bg\" (UID: \"2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409667 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da8ac6ff-99fb-4c50-a706-eec6c8987b04-config\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409684 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-config\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409700 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-audit-dir\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409721 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7e4947be-bb74-46c6-9a89-9c44dede6c7c-etcd-ca\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409737 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fs9l\" (UniqueName: \"kubernetes.io/projected/5c2eff22-d730-456b-870d-4138d9354577-kube-api-access-8fs9l\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.409952 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-audit-policies\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.410047 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4947be-bb74-46c6-9a89-9c44dede6c7c-config\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.410455 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7e4947be-bb74-46c6-9a89-9c44dede6c7c-etcd-service-ca\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.410656 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-audit-dir\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.411012 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.411710 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7e4947be-bb74-46c6-9a89-9c44dede6c7c-etcd-ca\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.411873 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-client-ca\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.412039 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/da8ac6ff-99fb-4c50-a706-eec6c8987b04-trusted-ca\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.412275 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da8ac6ff-99fb-4c50-a706-eec6c8987b04-config\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.412629 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.413320 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-config\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.414421 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-etcd-client\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.414435 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c2eff22-d730-456b-870d-4138d9354577-serving-cert\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.414811 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e4947be-bb74-46c6-9a89-9c44dede6c7c-serving-cert\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.415034 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-encryption-config\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.415060 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7e4947be-bb74-46c6-9a89-9c44dede6c7c-etcd-client\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.415209 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-serving-cert\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.415832 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da8ac6ff-99fb-4c50-a706-eec6c8987b04-serving-cert\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.416820 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f-metrics-tls\") pod \"dns-operator-744455d44c-vl6bg\" (UID: \"2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.427420 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.447499 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.453810 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.453854 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.453950 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.466665 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.486422 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.506611 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.528095 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.546964 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.567850 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.588335 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.607957 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.627242 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.633320 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/90685c8c-a90e-4821-9e19-40a677b71265-serviceca\") pod \"image-pruner-29560320-xlv4n\" (UID: \"90685c8c-a90e-4821-9e19-40a677b71265\") " pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.647717 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.687646 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.708331 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.733412 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.767620 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.787544 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.807954 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.827417 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.847893 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.868549 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.897419 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.907040 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.928433 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.947708 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.967315 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 16 00:08:10 crc kubenswrapper[4751]: I0316 00:08:10.988160 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.008808 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.027267 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.046775 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.067163 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.086933 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.107397 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.127496 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.148123 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.166943 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.186821 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.207153 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.227582 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.247468 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.267457 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.287424 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.305547 4751 request.go:700] Waited for 1.010815942s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dcatalog-operator-serving-cert&limit=500&resourceVersion=0 Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.307750 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.328306 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.347334 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.366597 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.387823 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.407857 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.433791 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.448148 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.453977 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.467756 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.493825 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.507522 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.527217 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.548215 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.572177 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.587499 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.627980 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.629834 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.648207 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.667862 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.688057 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.707309 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.726984 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.747006 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.767999 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.787502 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.807881 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.828040 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.865668 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t574r\" (UniqueName: \"kubernetes.io/projected/925c7b04-46f9-4ea0-91ce-4f098da81d33-kube-api-access-t574r\") pod \"openshift-apiserver-operator-796bbdcf4f-g5frw\" (UID: \"925c7b04-46f9-4ea0-91ce-4f098da81d33\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.887191 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.889898 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkn7w\" (UniqueName: \"kubernetes.io/projected/d455ed36-0071-411e-b306-237ca7d6695b-kube-api-access-nkn7w\") pod \"authentication-operator-69f744f599-dklns\" (UID: \"d455ed36-0071-411e-b306-237ca7d6695b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.907840 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.927970 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.947222 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.953759 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.967486 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.987178 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 16 00:08:11 crc kubenswrapper[4751]: I0316 00:08:11.990251 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.009241 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.027796 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.048769 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.074264 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.089253 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.108448 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.127957 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.149175 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.166482 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.187236 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.208446 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.225402 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw"] Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.227489 4751 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 16 00:08:12 crc kubenswrapper[4751]: W0316 00:08:12.234712 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod925c7b04_46f9_4ea0_91ce_4f098da81d33.slice/crio-c27f3664ae7b8427381a8543f6a8f9c2705e543fb5520eebd8bef3a6d453b7a7 WatchSource:0}: Error finding container c27f3664ae7b8427381a8543f6a8f9c2705e543fb5520eebd8bef3a6d453b7a7: Status 404 returned error can't find the container with id c27f3664ae7b8427381a8543f6a8f9c2705e543fb5520eebd8bef3a6d453b7a7 Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.246944 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.253054 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dklns"] Mar 16 00:08:12 crc kubenswrapper[4751]: W0316 00:08:12.261613 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd455ed36_0071_411e_b306_237ca7d6695b.slice/crio-32e7b4711001668324227d16a8fdff9fe8e6e29a0d89d67838a86cafdb2f1114 WatchSource:0}: Error finding container 32e7b4711001668324227d16a8fdff9fe8e6e29a0d89d67838a86cafdb2f1114: Status 404 returned error can't find the container with id 32e7b4711001668324227d16a8fdff9fe8e6e29a0d89d67838a86cafdb2f1114 Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.271176 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.287293 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.308173 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.325142 4751 request.go:700] Waited for 1.915206027s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.373684 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fs9l\" (UniqueName: \"kubernetes.io/projected/5c2eff22-d730-456b-870d-4138d9354577-kube-api-access-8fs9l\") pod \"route-controller-manager-6576b87f9c-wvjtz\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.423188 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k42kp\" (UniqueName: \"kubernetes.io/projected/90685c8c-a90e-4821-9e19-40a677b71265-kube-api-access-k42kp\") pod \"image-pruner-29560320-xlv4n\" (UID: \"90685c8c-a90e-4821-9e19-40a677b71265\") " pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.447145 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.469332 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.487280 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.507891 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534468 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/59c6efd1-70e9-45bc-8b87-94142678563d-metrics-tls\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534515 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534541 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-config\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534561 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnqqr\" (UniqueName: \"kubernetes.io/projected/a8fdc932-6670-45f5-9f69-c15eaae40bc4-kube-api-access-hnqqr\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534593 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b695c2c9-0b76-4e93-a427-76e157f001fb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534611 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-trusted-ca\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534642 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/74e4f428-25a3-45b9-959e-8a1eeb968cdf-node-pullsecrets\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534663 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74e4f428-25a3-45b9-959e-8a1eeb968cdf-serving-cert\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534683 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-config\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534703 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw4s5\" (UniqueName: \"kubernetes.io/projected/2394c21b-2d69-4698-8e58-d0198ec77558-kube-api-access-tw4s5\") pod \"downloads-7954f5f757-6k5cn\" (UID: \"2394c21b-2d69-4698-8e58-d0198ec77558\") " pod="openshift-console/downloads-7954f5f757-6k5cn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534722 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-oauth-config\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534743 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b695c2c9-0b76-4e93-a427-76e157f001fb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534763 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-audit\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534784 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534804 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534824 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwvwk\" (UniqueName: \"kubernetes.io/projected/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-kube-api-access-xwvwk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534845 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-trusted-ca-bundle\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534866 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d148ff3-a331-4c35-925c-3e3d0124a702-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534920 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-metrics-certs\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534939 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-console-config\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.534982 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk95r\" (UniqueName: \"kubernetes.io/projected/1c921c65-0fb5-4d15-b303-c9b7d0021931-kube-api-access-jk95r\") pod \"cluster-samples-operator-665b6dd947-qsvvx\" (UID: \"1c921c65-0fb5-4d15-b303-c9b7d0021931\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535011 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-image-import-ca\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535040 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535060 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59c6efd1-70e9-45bc-8b87-94142678563d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535080 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-machine-approver-tls\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535122 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj7k6\" (UniqueName: \"kubernetes.io/projected/a932e9e3-ca99-4af3-a9e6-8888e31be60d-kube-api-access-dj7k6\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535150 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvr6b\" (UniqueName: \"kubernetes.io/projected/74e4f428-25a3-45b9-959e-8a1eeb968cdf-kube-api-access-nvr6b\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535170 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-auth-proxy-config\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535201 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84j5b\" (UniqueName: \"kubernetes.io/projected/59c6efd1-70e9-45bc-8b87-94142678563d-kube-api-access-84j5b\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535221 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-serving-cert\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535240 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-oauth-serving-cert\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535260 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-certificates\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535279 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-bound-sa-token\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535299 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fdc932-6670-45f5-9f69-c15eaae40bc4-serving-cert\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535333 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/74e4f428-25a3-45b9-959e-8a1eeb968cdf-etcd-client\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535353 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-config\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535373 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-client-ca\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535393 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-stats-auth\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535431 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535478 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/74e4f428-25a3-45b9-959e-8a1eeb968cdf-audit-dir\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535499 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-default-certificate\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535518 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ngn4\" (UniqueName: \"kubernetes.io/projected/cfe3c553-b411-4074-b83a-c8eba2518089-kube-api-access-6ngn4\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535538 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a932e9e3-ca99-4af3-a9e6-8888e31be60d-config\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535560 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a932e9e3-ca99-4af3-a9e6-8888e31be60d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535588 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-tls\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535617 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d148ff3-a331-4c35-925c-3e3d0124a702-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535638 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a792c668-ee31-4bd8-a122-37bf8b48f37f-service-ca-bundle\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535659 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfgxp\" (UniqueName: \"kubernetes.io/projected/a792c668-ee31-4bd8-a122-37bf8b48f37f-kube-api-access-nfgxp\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535682 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2fgg\" (UniqueName: \"kubernetes.io/projected/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-kube-api-access-l2fgg\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535704 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a932e9e3-ca99-4af3-a9e6-8888e31be60d-images\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535724 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-config\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535750 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-serving-cert\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535770 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535789 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-service-ca\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535810 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn2cx\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-kube-api-access-dn2cx\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535830 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/74e4f428-25a3-45b9-959e-8a1eeb968cdf-encryption-config\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535848 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59c6efd1-70e9-45bc-8b87-94142678563d-trusted-ca\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.535876 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.536075 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c921c65-0fb5-4d15-b303-c9b7d0021931-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qsvvx\" (UID: \"1c921c65-0fb5-4d15-b303-c9b7d0021931\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.536118 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d148ff3-a331-4c35-925c-3e3d0124a702-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.536139 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdgj4\" (UniqueName: \"kubernetes.io/projected/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-kube-api-access-cdgj4\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.536179 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-etcd-serving-ca\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.536200 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:12 crc kubenswrapper[4751]: E0316 00:08:12.537534 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.037517175 +0000 UTC m=+121.306849475 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.567270 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.588076 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.606742 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.628362 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.636670 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:12 crc kubenswrapper[4751]: E0316 00:08:12.636847 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.136810927 +0000 UTC m=+121.406143227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.636964 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a932e9e3-ca99-4af3-a9e6-8888e31be60d-config\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637022 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-dir\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637079 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a932e9e3-ca99-4af3-a9e6-8888e31be60d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637151 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9118707-2746-44ed-bc4d-004312b53b5d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637203 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-csi-data-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637236 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8lhg\" (UniqueName: \"kubernetes.io/projected/8545bd4e-a993-4f46-9d64-963a5f685a2d-kube-api-access-m8lhg\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637289 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-tls\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637340 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d148ff3-a331-4c35-925c-3e3d0124a702-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637371 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmdpq\" (UniqueName: \"kubernetes.io/projected/adb49d8b-968b-4fb1-8464-17853a08dda8-kube-api-access-mmdpq\") pod \"ingress-canary-nsnh2\" (UID: \"adb49d8b-968b-4fb1-8464-17853a08dda8\") " pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637402 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-policies\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637483 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a932e9e3-ca99-4af3-a9e6-8888e31be60d-images\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637541 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh7cp\" (UniqueName: \"kubernetes.io/projected/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-kube-api-access-bh7cp\") pod \"package-server-manager-789f6589d5-xp482\" (UID: \"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637581 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e707425-3377-458e-a331-6af00d1e989c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637618 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9c56ee92-a6f5-46ec-8076-157268b354c7-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637655 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637690 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-service-ca\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637731 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn2cx\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-kube-api-access-dn2cx\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637767 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637815 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637849 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9c56ee92-a6f5-46ec-8076-157268b354c7-ready\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637881 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xp482\" (UID: \"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.637932 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b60da8df-4808-44ec-8e43-bc5e7b56079d-config\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638022 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b6d655c6-0506-41b7-b032-842ff9630e41-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638085 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638213 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-etcd-serving-ca\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638265 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638342 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-config\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638569 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638578 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shm2f\" (UniqueName: \"kubernetes.io/projected/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-kube-api-access-shm2f\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638751 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b695c2c9-0b76-4e93-a427-76e157f001fb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638790 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-trusted-ca\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638818 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638855 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74e4f428-25a3-45b9-959e-8a1eeb968cdf-serving-cert\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638883 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69gt4\" (UniqueName: \"kubernetes.io/projected/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-kube-api-access-69gt4\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638906 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-profile-collector-cert\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638933 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-audit\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638965 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-serving-cert\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.638999 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639031 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1ee3303-f745-4058-8927-046b07740c33-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6qwkx\" (UID: \"c1ee3303-f745-4058-8927-046b07740c33\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639130 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcpx7\" (UniqueName: \"kubernetes.io/projected/0f24b5d5-6e99-4e6a-98b9-ea489c774893-kube-api-access-rcpx7\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639253 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639297 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639330 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-trusted-ca-bundle\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639386 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb49d\" (UniqueName: \"kubernetes.io/projected/c9118707-2746-44ed-bc4d-004312b53b5d-kube-api-access-pb49d\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639422 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-mountpoint-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639458 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639527 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-webhook-cert\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639539 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b695c2c9-0b76-4e93-a427-76e157f001fb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639601 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639687 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6d655c6-0506-41b7-b032-842ff9630e41-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639725 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639785 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-image-import-ca\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639826 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60da8df-4808-44ec-8e43-bc5e7b56079d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639884 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-machine-approver-tls\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639918 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj7k6\" (UniqueName: \"kubernetes.io/projected/a932e9e3-ca99-4af3-a9e6-8888e31be60d-kube-api-access-dj7k6\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639957 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76wnr\" (UniqueName: \"kubernetes.io/projected/c2676895-4d0e-4099-9d49-abd723369982-kube-api-access-76wnr\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.639990 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-srv-cert\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640027 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-auth-proxy-config\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640185 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rprdx\" (UniqueName: \"kubernetes.io/projected/c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05-kube-api-access-rprdx\") pod \"auto-csr-approver-29560328-lt56z\" (UID: \"c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05\") " pod="openshift-infra/auto-csr-approver-29560328-lt56z" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640225 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9c56ee92-a6f5-46ec-8076-157268b354c7-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640264 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-oauth-serving-cert\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640316 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640369 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-bound-sa-token\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640420 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fdc932-6670-45f5-9f69-c15eaae40bc4-serving-cert\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640505 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/74e4f428-25a3-45b9-959e-8a1eeb968cdf-etcd-client\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640540 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-certs\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640574 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vn22\" (UniqueName: \"kubernetes.io/projected/b6d655c6-0506-41b7-b032-842ff9630e41-kube-api-access-5vn22\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640607 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640645 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-stats-auth\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640713 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-node-bootstrap-token\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640779 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-config-volume\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640820 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6d655c6-0506-41b7-b032-842ff9630e41-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640853 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d8ht\" (UniqueName: \"kubernetes.io/projected/9c56ee92-a6f5-46ec-8076-157268b354c7-kube-api-access-8d8ht\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640915 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-default-certificate\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640958 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ngn4\" (UniqueName: \"kubernetes.io/projected/cfe3c553-b411-4074-b83a-c8eba2518089-kube-api-access-6ngn4\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.640993 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/270222df-3862-47e2-a363-c69546415e18-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641029 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e707425-3377-458e-a331-6af00d1e989c-proxy-tls\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641060 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsfc2\" (UniqueName: \"kubernetes.io/projected/1e707425-3377-458e-a331-6af00d1e989c-kube-api-access-gsfc2\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641155 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-registration-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641194 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641374 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a792c668-ee31-4bd8-a122-37bf8b48f37f-service-ca-bundle\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641547 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfgxp\" (UniqueName: \"kubernetes.io/projected/a792c668-ee31-4bd8-a122-37bf8b48f37f-kube-api-access-nfgxp\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641617 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-srv-cert\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641699 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2fgg\" (UniqueName: \"kubernetes.io/projected/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-kube-api-access-l2fgg\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641735 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9118707-2746-44ed-bc4d-004312b53b5d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641772 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-config\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641809 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-serving-cert\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641843 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2676895-4d0e-4099-9d49-abd723369982-config-volume\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641880 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv9n2\" (UniqueName: \"kubernetes.io/projected/270222df-3862-47e2-a363-c69546415e18-kube-api-access-vv9n2\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.641916 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1e707425-3377-458e-a331-6af00d1e989c-images\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642016 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/74e4f428-25a3-45b9-959e-8a1eeb968cdf-encryption-config\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642084 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59c6efd1-70e9-45bc-8b87-94142678563d-trusted-ca\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642198 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk6sg\" (UniqueName: \"kubernetes.io/projected/69dee6c1-1c2f-4d49-8955-ca4226b4306e-kube-api-access-fk6sg\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642259 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdgj4\" (UniqueName: \"kubernetes.io/projected/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-kube-api-access-cdgj4\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642333 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c921c65-0fb5-4d15-b303-c9b7d0021931-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qsvvx\" (UID: \"1c921c65-0fb5-4d15-b303-c9b7d0021931\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642383 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d148ff3-a331-4c35-925c-3e3d0124a702-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642427 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/adb49d8b-968b-4fb1-8464-17853a08dda8-cert\") pod \"ingress-canary-nsnh2\" (UID: \"adb49d8b-968b-4fb1-8464-17853a08dda8\") " pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642661 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmptm\" (UniqueName: \"kubernetes.io/projected/b1a71ab9-7855-40cf-9670-86381e49c4bc-kube-api-access-pmptm\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642732 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.642783 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/59c6efd1-70e9-45bc-8b87-94142678563d-metrics-tls\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643244 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnqqr\" (UniqueName: \"kubernetes.io/projected/a8fdc932-6670-45f5-9f69-c15eaae40bc4-kube-api-access-hnqqr\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643395 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-key\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643560 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/74e4f428-25a3-45b9-959e-8a1eeb968cdf-node-pullsecrets\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643642 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/74e4f428-25a3-45b9-959e-8a1eeb968cdf-node-pullsecrets\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643649 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-config\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643723 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw4s5\" (UniqueName: \"kubernetes.io/projected/2394c21b-2d69-4698-8e58-d0198ec77558-kube-api-access-tw4s5\") pod \"downloads-7954f5f757-6k5cn\" (UID: \"2394c21b-2d69-4698-8e58-d0198ec77558\") " pod="openshift-console/downloads-7954f5f757-6k5cn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643762 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-oauth-config\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643798 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-plugins-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643834 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcp7q\" (UniqueName: \"kubernetes.io/projected/c0f92561-e30f-4e32-80d4-b463fd87de66-kube-api-access-mcp7q\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643874 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b695c2c9-0b76-4e93-a427-76e157f001fb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643908 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-metrics-tls\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643943 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwvwk\" (UniqueName: \"kubernetes.io/projected/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-kube-api-access-xwvwk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.643979 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-socket-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644012 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2676895-4d0e-4099-9d49-abd723369982-secret-volume\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644052 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d148ff3-a331-4c35-925c-3e3d0124a702-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644132 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-console-config\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644203 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-metrics-certs\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644258 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk95r\" (UniqueName: \"kubernetes.io/projected/1c921c65-0fb5-4d15-b303-c9b7d0021931-kube-api-access-jk95r\") pod \"cluster-samples-operator-665b6dd947-qsvvx\" (UID: \"1c921c65-0fb5-4d15-b303-c9b7d0021931\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644296 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksrcg\" (UniqueName: \"kubernetes.io/projected/5620fac3-59f5-49e4-9125-501353fbf836-kube-api-access-ksrcg\") pod \"control-plane-machine-set-operator-78cbb6b69f-w4zmm\" (UID: \"5620fac3-59f5-49e4-9125-501353fbf836\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644482 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8545bd4e-a993-4f46-9d64-963a5f685a2d-tmpfs\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644517 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644548 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59c6efd1-70e9-45bc-8b87-94142678563d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644580 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8bzf\" (UniqueName: \"kubernetes.io/projected/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-kube-api-access-h8bzf\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644614 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79vjv\" (UniqueName: \"kubernetes.io/projected/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-kube-api-access-79vjv\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644644 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-cabundle\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644674 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644710 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvr6b\" (UniqueName: \"kubernetes.io/projected/74e4f428-25a3-45b9-959e-8a1eeb968cdf-kube-api-access-nvr6b\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644743 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/270222df-3862-47e2-a363-c69546415e18-proxy-tls\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.644826 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjxhz\" (UniqueName: \"kubernetes.io/projected/2b4bf065-6c73-43a0-b02f-a03188a64add-kube-api-access-xjxhz\") pod \"migrator-59844c95c7-g9fxd\" (UID: \"2b4bf065-6c73-43a0-b02f-a03188a64add\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645009 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfhrv\" (UniqueName: \"kubernetes.io/projected/e55dc627-34bc-4853-ba79-146754ae58b5-kube-api-access-mfhrv\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645191 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5620fac3-59f5-49e4-9125-501353fbf836-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-w4zmm\" (UID: \"5620fac3-59f5-49e4-9125-501353fbf836\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645288 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84j5b\" (UniqueName: \"kubernetes.io/projected/59c6efd1-70e9-45bc-8b87-94142678563d-kube-api-access-84j5b\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645365 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj5tl\" (UniqueName: \"kubernetes.io/projected/c1ee3303-f745-4058-8927-046b07740c33-kube-api-access-vj5tl\") pod \"multus-admission-controller-857f4d67dd-6qwkx\" (UID: \"c1ee3303-f745-4058-8927-046b07740c33\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645501 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645553 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-serving-cert\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645597 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-certificates\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645641 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-apiservice-cert\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645677 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645725 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-config\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645762 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-client-ca\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645800 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645870 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645921 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/74e4f428-25a3-45b9-959e-8a1eeb968cdf-audit-dir\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.645969 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-config\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.646012 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60da8df-4808-44ec-8e43-bc5e7b56079d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.646335 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/74e4f428-25a3-45b9-959e-8a1eeb968cdf-audit-dir\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: E0316 00:08:12.646354 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.146332613 +0000 UTC m=+121.415664943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.648637 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-certificates\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.649068 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.667026 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.677278 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.687455 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.709457 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.719777 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.747272 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:12 crc kubenswrapper[4751]: E0316 00:08:12.747574 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.247506222 +0000 UTC m=+121.516838542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748060 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-apiservice-cert\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748144 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748217 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748280 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748325 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-config\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.747944 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748375 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60da8df-4808-44ec-8e43-bc5e7b56079d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748736 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-dir\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748805 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9118707-2746-44ed-bc4d-004312b53b5d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748844 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-csi-data-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748879 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8lhg\" (UniqueName: \"kubernetes.io/projected/8545bd4e-a993-4f46-9d64-963a5f685a2d-kube-api-access-m8lhg\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748938 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmdpq\" (UniqueName: \"kubernetes.io/projected/adb49d8b-968b-4fb1-8464-17853a08dda8-kube-api-access-mmdpq\") pod \"ingress-canary-nsnh2\" (UID: \"adb49d8b-968b-4fb1-8464-17853a08dda8\") " pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.748971 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-policies\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749015 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh7cp\" (UniqueName: \"kubernetes.io/projected/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-kube-api-access-bh7cp\") pod \"package-server-manager-789f6589d5-xp482\" (UID: \"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749052 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e707425-3377-458e-a331-6af00d1e989c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749098 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9c56ee92-a6f5-46ec-8076-157268b354c7-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: E0316 00:08:12.749153 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.249093819 +0000 UTC m=+121.518426149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749214 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749242 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9c56ee92-a6f5-46ec-8076-157268b354c7-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749278 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9c56ee92-a6f5-46ec-8076-157268b354c7-ready\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749305 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-dir\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749318 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xp482\" (UID: \"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749353 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b60da8df-4808-44ec-8e43-bc5e7b56079d-config\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749387 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749453 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b6d655c6-0506-41b7-b032-842ff9630e41-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749469 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-csi-data-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749520 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shm2f\" (UniqueName: \"kubernetes.io/projected/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-kube-api-access-shm2f\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749565 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749610 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69gt4\" (UniqueName: \"kubernetes.io/projected/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-kube-api-access-69gt4\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749644 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-profile-collector-cert\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749690 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-serving-cert\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749721 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749755 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcpx7\" (UniqueName: \"kubernetes.io/projected/0f24b5d5-6e99-4e6a-98b9-ea489c774893-kube-api-access-rcpx7\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749825 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1ee3303-f745-4058-8927-046b07740c33-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6qwkx\" (UID: \"c1ee3303-f745-4058-8927-046b07740c33\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749861 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb49d\" (UniqueName: \"kubernetes.io/projected/c9118707-2746-44ed-bc4d-004312b53b5d-kube-api-access-pb49d\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749894 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-mountpoint-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749930 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.749968 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-webhook-cert\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750003 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750041 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6d655c6-0506-41b7-b032-842ff9630e41-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750099 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750235 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76wnr\" (UniqueName: \"kubernetes.io/projected/c2676895-4d0e-4099-9d49-abd723369982-kube-api-access-76wnr\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750269 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-srv-cert\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750300 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60da8df-4808-44ec-8e43-bc5e7b56079d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750348 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rprdx\" (UniqueName: \"kubernetes.io/projected/c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05-kube-api-access-rprdx\") pod \"auto-csr-approver-29560328-lt56z\" (UID: \"c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05\") " pod="openshift-infra/auto-csr-approver-29560328-lt56z" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750383 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9c56ee92-a6f5-46ec-8076-157268b354c7-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750426 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750487 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-certs\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750520 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vn22\" (UniqueName: \"kubernetes.io/projected/b6d655c6-0506-41b7-b032-842ff9630e41-kube-api-access-5vn22\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750568 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-node-bootstrap-token\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750600 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750630 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e707425-3377-458e-a331-6af00d1e989c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750640 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6d655c6-0506-41b7-b032-842ff9630e41-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750723 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d8ht\" (UniqueName: \"kubernetes.io/projected/9c56ee92-a6f5-46ec-8076-157268b354c7-kube-api-access-8d8ht\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750784 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/270222df-3862-47e2-a363-c69546415e18-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750818 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-config-volume\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750852 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e707425-3377-458e-a331-6af00d1e989c-proxy-tls\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750885 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsfc2\" (UniqueName: \"kubernetes.io/projected/1e707425-3377-458e-a331-6af00d1e989c-kube-api-access-gsfc2\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750921 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-registration-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750974 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.751028 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-srv-cert\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.751074 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9118707-2746-44ed-bc4d-004312b53b5d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.750593 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9c56ee92-a6f5-46ec-8076-157268b354c7-ready\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.752090 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2676895-4d0e-4099-9d49-abd723369982-config-volume\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.752185 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1e707425-3377-458e-a331-6af00d1e989c-images\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.752245 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv9n2\" (UniqueName: \"kubernetes.io/projected/270222df-3862-47e2-a363-c69546415e18-kube-api-access-vv9n2\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.752284 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk6sg\" (UniqueName: \"kubernetes.io/projected/69dee6c1-1c2f-4d49-8955-ca4226b4306e-kube-api-access-fk6sg\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.752384 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-mountpoint-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.752404 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmptm\" (UniqueName: \"kubernetes.io/projected/b1a71ab9-7855-40cf-9670-86381e49c4bc-kube-api-access-pmptm\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.752820 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-registration-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.752893 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/adb49d8b-968b-4fb1-8464-17853a08dda8-cert\") pod \"ingress-canary-nsnh2\" (UID: \"adb49d8b-968b-4fb1-8464-17853a08dda8\") " pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753069 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-key\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753194 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-plugins-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753277 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcp7q\" (UniqueName: \"kubernetes.io/projected/c0f92561-e30f-4e32-80d4-b463fd87de66-kube-api-access-mcp7q\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753357 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/270222df-3862-47e2-a363-c69546415e18-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753374 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-metrics-tls\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753557 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-socket-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753652 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2676895-4d0e-4099-9d49-abd723369982-secret-volume\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753835 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-plugins-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753846 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksrcg\" (UniqueName: \"kubernetes.io/projected/5620fac3-59f5-49e4-9125-501353fbf836-kube-api-access-ksrcg\") pod \"control-plane-machine-set-operator-78cbb6b69f-w4zmm\" (UID: \"5620fac3-59f5-49e4-9125-501353fbf836\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.753966 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8bzf\" (UniqueName: \"kubernetes.io/projected/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-kube-api-access-h8bzf\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.754006 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79vjv\" (UniqueName: \"kubernetes.io/projected/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-kube-api-access-79vjv\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.754312 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b1a71ab9-7855-40cf-9670-86381e49c4bc-socket-dir\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.754559 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-cabundle\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.754638 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8545bd4e-a993-4f46-9d64-963a5f685a2d-tmpfs\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.754694 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.754772 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/270222df-3862-47e2-a363-c69546415e18-proxy-tls\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.754823 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjxhz\" (UniqueName: \"kubernetes.io/projected/2b4bf065-6c73-43a0-b02f-a03188a64add-kube-api-access-xjxhz\") pod \"migrator-59844c95c7-g9fxd\" (UID: \"2b4bf065-6c73-43a0-b02f-a03188a64add\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.754877 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfhrv\" (UniqueName: \"kubernetes.io/projected/e55dc627-34bc-4853-ba79-146754ae58b5-kube-api-access-mfhrv\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.754929 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5620fac3-59f5-49e4-9125-501353fbf836-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-w4zmm\" (UID: \"5620fac3-59f5-49e4-9125-501353fbf836\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.755007 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj5tl\" (UniqueName: \"kubernetes.io/projected/c1ee3303-f745-4058-8927-046b07740c33-kube-api-access-vj5tl\") pod \"multus-admission-controller-857f4d67dd-6qwkx\" (UID: \"c1ee3303-f745-4058-8927-046b07740c33\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.755074 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.755290 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8545bd4e-a993-4f46-9d64-963a5f685a2d-tmpfs\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.760016 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn2sw\" (UniqueName: \"kubernetes.io/projected/2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f-kube-api-access-cn2sw\") pod \"dns-operator-744455d44c-vl6bg\" (UID: \"2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.769809 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.778326 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmgbb\" (UniqueName: \"kubernetes.io/projected/8e87f4be-4c72-4d7d-a2b7-4584d11be3c8-kube-api-access-fmgbb\") pod \"apiserver-7bbb656c7d-tj4rk\" (UID: \"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.788391 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.800420 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a932e9e3-ca99-4af3-a9e6-8888e31be60d-config\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.808178 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.823146 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a932e9e3-ca99-4af3-a9e6-8888e31be60d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.830941 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.842505 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-tls\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.857976 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:12 crc kubenswrapper[4751]: E0316 00:08:12.859764 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.359722081 +0000 UTC m=+121.629054381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.860411 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: E0316 00:08:12.860907 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.360890786 +0000 UTC m=+121.630223086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.871859 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.882541 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a932e9e3-ca99-4af3-a9e6-8888e31be60d-images\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.907655 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.909591 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn2cx\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-kube-api-access-dn2cx\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.911273 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.930930 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.931278 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz"] Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.939757 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-service-ca\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.947025 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.949820 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-etcd-serving-ca\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.961967 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:12 crc kubenswrapper[4751]: E0316 00:08:12.962092 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.462066805 +0000 UTC m=+121.731399095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.962926 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:12 crc kubenswrapper[4751]: E0316 00:08:12.963298 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.463281861 +0000 UTC m=+121.732614161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.983708 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29560320-xlv4n"] Mar 16 00:08:12 crc kubenswrapper[4751]: W0316 00:08:12.984848 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90685c8c_a90e_4821_9e19_40a677b71265.slice/crio-125254adf9db91e886bfc5390613cfe18d8ffd0c40d1ce46fe71928f22fedce1 WatchSource:0}: Error finding container 125254adf9db91e886bfc5390613cfe18d8ffd0c40d1ce46fe71928f22fedce1: Status 404 returned error can't find the container with id 125254adf9db91e886bfc5390613cfe18d8ffd0c40d1ce46fe71928f22fedce1 Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.988417 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 16 00:08:12 crc kubenswrapper[4751]: I0316 00:08:12.989718 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-config\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.014556 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.020211 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-trusted-ca\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.023427 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6d655c6-0506-41b7-b032-842ff9630e41-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.026718 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.036719 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74e4f428-25a3-45b9-959e-8a1eeb968cdf-serving-cert\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.049301 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.050920 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-audit\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.064637 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.064752 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.564723108 +0000 UTC m=+121.834055408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.066261 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.066932 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.566866653 +0000 UTC m=+121.836198983 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.078651 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.081739 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-trusted-ca-bundle\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.087365 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.092998 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.114533 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.121561 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.127301 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.131489 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-image-import-ca\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.143357 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" event={"ID":"d455ed36-0071-411e-b306-237ca7d6695b","Type":"ContainerStarted","Data":"66025a24ded3895431fbee1a7c231cc9e9036d3973b4f3457ac132efe97e8511"} Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.143414 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" event={"ID":"d455ed36-0071-411e-b306-237ca7d6695b","Type":"ContainerStarted","Data":"32e7b4711001668324227d16a8fdff9fe8e6e29a0d89d67838a86cafdb2f1114"} Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.145187 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" event={"ID":"925c7b04-46f9-4ea0-91ce-4f098da81d33","Type":"ContainerStarted","Data":"9da917c208ed2991ad88cc4740ab66f80e611a2b68da76d3fb34976074cecf86"} Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.145246 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" event={"ID":"925c7b04-46f9-4ea0-91ce-4f098da81d33","Type":"ContainerStarted","Data":"c27f3664ae7b8427381a8543f6a8f9c2705e543fb5520eebd8bef3a6d453b7a7"} Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.146803 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29560320-xlv4n" event={"ID":"90685c8c-a90e-4821-9e19-40a677b71265","Type":"ContainerStarted","Data":"d7c7bd7974f3e6c328436d2e50ae6f62328e5c16c85ffcbc211729a42c91093e"} Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.146946 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29560320-xlv4n" event={"ID":"90685c8c-a90e-4821-9e19-40a677b71265","Type":"ContainerStarted","Data":"125254adf9db91e886bfc5390613cfe18d8ffd0c40d1ce46fe71928f22fedce1"} Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.147479 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.151080 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" event={"ID":"5c2eff22-d730-456b-870d-4138d9354577","Type":"ContainerStarted","Data":"cc11ef3da5a2375d5ec5d6fe7bc35382471a765532ece0e87f87e9afb820c437"} Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.151138 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" event={"ID":"5c2eff22-d730-456b-870d-4138d9354577","Type":"ContainerStarted","Data":"1a783ea48f6fb8e6a6b842fa7de6aaa0eb69c5cd0608f9d72bf2f0a2cc8b79c9"} Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.151576 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.153615 4751 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-wvjtz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.153772 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" podUID="5c2eff22-d730-456b-870d-4138d9354577" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.154136 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-machine-approver-tls\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.167130 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.167394 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.667350401 +0000 UTC m=+121.936682741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.168153 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.168512 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.168953 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.668937898 +0000 UTC m=+121.938270198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.171311 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-auth-proxy-config\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.225328 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-bound-sa-token\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.227022 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.231198 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-oauth-serving-cert\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.247685 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.254365 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fdc932-6670-45f5-9f69-c15eaae40bc4-serving-cert\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.267914 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.272244 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.272451 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.772422957 +0000 UTC m=+122.041755257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.273207 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.275660 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.775638773 +0000 UTC m=+122.044971073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.276083 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/74e4f428-25a3-45b9-959e-8a1eeb968cdf-etcd-client\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.287278 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.296702 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-stats-auth\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.306640 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.315551 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-default-certificate\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.352609 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.354721 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a792c668-ee31-4bd8-a122-37bf8b48f37f-service-ca-bundle\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.374776 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.375016 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.874988437 +0000 UTC m=+122.144320737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.375426 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.375944 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.875914485 +0000 UTC m=+122.145246805 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.406930 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.414432 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-config\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.428697 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.436709 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-serving-cert\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.447623 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.456631 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/74e4f428-25a3-45b9-959e-8a1eeb968cdf-encryption-config\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.477343 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.477343 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.477446 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.977418994 +0000 UTC m=+122.246751324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.478567 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.479005 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:13.978990081 +0000 UTC m=+122.248322381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.485476 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59c6efd1-70e9-45bc-8b87-94142678563d-trusted-ca\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.507276 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.516691 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c921c65-0fb5-4d15-b303-c9b7d0021931-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qsvvx\" (UID: \"1c921c65-0fb5-4d15-b303-c9b7d0021931\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.527737 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.538290 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d148ff3-a331-4c35-925c-3e3d0124a702-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.548079 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.553385 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.567268 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.577435 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/59c6efd1-70e9-45bc-8b87-94142678563d-metrics-tls\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.578966 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.579168 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.079143019 +0000 UTC m=+122.348475349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.606750 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.614468 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-config\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.644912 4751 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.644952 4751 secret.go:188] Couldn't get secret openshift-console/console-oauth-config: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.644978 4751 secret.go:188] Couldn't get secret openshift-image-registry/installation-pull-secrets: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.644930 4751 secret.go:188] Couldn't get secret openshift-ingress/router-metrics-certs-default: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.645040 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-trusted-ca-bundle podName:74e4f428-25a3-45b9-959e-8a1eeb968cdf nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.144988849 +0000 UTC m=+122.414321149 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-trusted-ca-bundle") pod "apiserver-76f77b778f-9mrng" (UID: "74e4f428-25a3-45b9-959e-8a1eeb968cdf") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.645034 4751 configmap.go:193] Couldn't get configMap openshift-console/console-config: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.645125 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-metrics-certs podName:a792c668-ee31-4bd8-a122-37bf8b48f37f nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.145055541 +0000 UTC m=+122.414387821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-metrics-certs") pod "router-default-5444994796-jpvk6" (UID: "a792c668-ee31-4bd8-a122-37bf8b48f37f") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.644914 4751 configmap.go:193] Couldn't get configMap openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.645151 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-oauth-config podName:cfe3c553-b411-4074-b83a-c8eba2518089 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.145144244 +0000 UTC m=+122.414476534 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-oauth-config" (UniqueName: "kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-oauth-config") pod "console-f9d7485db-bd5b9" (UID: "cfe3c553-b411-4074-b83a-c8eba2518089") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.645155 4751 request.go:700] Waited for 1.00110496s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/secrets?fieldSelector=metadata.name%3Dconsole-oauth-config&resourceVersion=27035 Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.645266 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b695c2c9-0b76-4e93-a427-76e157f001fb-installation-pull-secrets podName:b695c2c9-0b76-4e93-a427-76e157f001fb nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.145238057 +0000 UTC m=+122.414570447 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "installation-pull-secrets" (UniqueName: "kubernetes.io/secret/b695c2c9-0b76-4e93-a427-76e157f001fb-installation-pull-secrets") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.645296 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-console-config podName:cfe3c553-b411-4074-b83a-c8eba2518089 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.145283448 +0000 UTC m=+122.414615888 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-config" (UniqueName: "kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-console-config") pod "console-f9d7485db-bd5b9" (UID: "cfe3c553-b411-4074-b83a-c8eba2518089") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.645320 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2d148ff3-a331-4c35-925c-3e3d0124a702-config podName:2d148ff3-a331-4c35-925c-3e3d0124a702 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.145310999 +0000 UTC m=+122.414643399 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/2d148ff3-a331-4c35-925c-3e3d0124a702-config") pod "openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" (UID: "2d148ff3-a331-4c35-925c-3e3d0124a702") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.645673 4751 secret.go:188] Couldn't get secret openshift-console/console-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.645725 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-serving-cert podName:cfe3c553-b411-4074-b83a-c8eba2518089 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.145711381 +0000 UTC m=+122.415043751 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-serving-cert" (UniqueName: "kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-serving-cert") pod "console-f9d7485db-bd5b9" (UID: "cfe3c553-b411-4074-b83a-c8eba2518089") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.646188 4751 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.646238 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-client-ca podName:a8fdc932-6670-45f5-9f69-c15eaae40bc4 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.146226956 +0000 UTC m=+122.415559336 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-client-ca") pod "controller-manager-879f6c89f-mrm4x" (UID: "a8fdc932-6670-45f5-9f69-c15eaae40bc4") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.646391 4751 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.646527 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-config podName:a8fdc932-6670-45f5-9f69-c15eaae40bc4 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.146508355 +0000 UTC m=+122.415840715 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-config") pod "controller-manager-879f6c89f-mrm4x" (UID: "a8fdc932-6670-45f5-9f69-c15eaae40bc4") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.647890 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.652565 4751 projected.go:288] Couldn't get configMap openshift-console-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.652595 4751 projected.go:194] Error preparing data for projected volume kube-api-access-8bzhm for pod openshift-console-operator/console-operator-58897d9998-njjzb: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.652653 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/da8ac6ff-99fb-4c50-a706-eec6c8987b04-kube-api-access-8bzhm podName:da8ac6ff-99fb-4c50-a706-eec6c8987b04 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.152637238 +0000 UTC m=+122.421969538 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-8bzhm" (UniqueName: "kubernetes.io/projected/da8ac6ff-99fb-4c50-a706-eec6c8987b04-kube-api-access-8bzhm") pod "console-operator-58897d9998-njjzb" (UID: "da8ac6ff-99fb-4c50-a706-eec6c8987b04") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.667975 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.681567 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.682256 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.182226914 +0000 UTC m=+122.451559214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.690259 4751 projected.go:288] Couldn't get configMap openshift-etcd-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.690521 4751 projected.go:194] Error preparing data for projected volume kube-api-access-l8kxm for pod openshift-etcd-operator/etcd-operator-b45778765-jszhl: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.690644 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7e4947be-bb74-46c6-9a89-9c44dede6c7c-kube-api-access-l8kxm podName:7e4947be-bb74-46c6-9a89-9c44dede6c7c nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.190622416 +0000 UTC m=+122.459954706 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-l8kxm" (UniqueName: "kubernetes.io/projected/7e4947be-bb74-46c6-9a89-9c44dede6c7c-kube-api-access-l8kxm") pod "etcd-operator-b45778765-jszhl" (UID: "7e4947be-bb74-46c6-9a89-9c44dede6c7c") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.706223 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.727038 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.747951 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.748425 4751 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-error: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.748476 4751 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.748526 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-error podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.248497008 +0000 UTC m=+122.517829318 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-error" (UniqueName: "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-error") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.748586 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-apiservice-cert podName:8545bd4e-a993-4f46-9d64-963a5f685a2d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.24855994 +0000 UTC m=+122.517892310 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-apiservice-cert") pod "packageserver-d55dfcdfc-rl4gt" (UID: "8545bd4e-a993-4f46-9d64-963a5f685a2d") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.748785 4751 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.748938 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-provider-selection podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.24890131 +0000 UTC m=+122.518233620 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.749427 4751 configmap.go:193] Couldn't get configMap openshift-kube-storage-version-migrator-operator/config: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.749563 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c9118707-2746-44ed-bc4d-004312b53b5d-config podName:c9118707-2746-44ed-bc4d-004312b53b5d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.24954796 +0000 UTC m=+122.518880260 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/c9118707-2746-44ed-bc4d-004312b53b5d-config") pod "kube-storage-version-migrator-operator-b67b599dd-c9jpp" (UID: "c9118707-2746-44ed-bc4d-004312b53b5d") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.749651 4751 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.749728 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-config podName:488dbfc7-cfc6-4c83-98d4-b2bc9f16969d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.249714375 +0000 UTC m=+122.519046685 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-config") pod "service-ca-operator-777779d784-znlfm" (UID: "488dbfc7-cfc6-4c83-98d4-b2bc9f16969d") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.749852 4751 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.749984 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-policies podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.249968302 +0000 UTC m=+122.519300602 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-policies") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.750019 4751 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-router-certs: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.750198 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-router-certs podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.250187329 +0000 UTC m=+122.519519629 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-router-certs" (UniqueName: "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-router-certs") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751661 4751 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751758 4751 configmap.go:193] Couldn't get configMap openshift-kube-controller-manager-operator/kube-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751763 4751 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751791 4751 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751819 4751 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751821 4751 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751845 4751 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751869 4751 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751769 4751 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751832 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b60da8df-4808-44ec-8e43-bc5e7b56079d-config podName:b60da8df-4808-44ec-8e43-bc5e7b56079d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.251819348 +0000 UTC m=+122.521151648 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/b60da8df-4808-44ec-8e43-bc5e7b56079d-config") pod "kube-controller-manager-operator-78b949d7b-wdvsd" (UID: "b60da8df-4808-44ec-8e43-bc5e7b56079d") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751934 4751 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751927 4751 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751698 4751 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-session: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751962 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-profile-collector-cert podName:a3f6093c-dd51-45d1-9d18-0a80bac4fdb1 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.251940201 +0000 UTC m=+122.521272481 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-profile-collector-cert") pod "catalog-operator-68c6474976-rfk2f" (UID: "a3f6093c-dd51-45d1-9d18-0a80bac4fdb1") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752039 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1ee3303-f745-4058-8927-046b07740c33-webhook-certs podName:c1ee3303-f745-4058-8927-046b07740c33 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252020324 +0000 UTC m=+122.521352624 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1ee3303-f745-4058-8927-046b07740c33-webhook-certs") pod "multus-admission-controller-857f4d67dd-6qwkx" (UID: "c1ee3303-f745-4058-8927-046b07740c33") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752060 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-trusted-ca podName:69dee6c1-1c2f-4d49-8955-ca4226b4306e nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252049705 +0000 UTC m=+122.521382005 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-trusted-ca") pod "marketplace-operator-79b997595-sph8v" (UID: "69dee6c1-1c2f-4d49-8955-ca4226b4306e") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.751789 4751 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752119 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-certs podName:0f24b5d5-6e99-4e6a-98b9-ea489c774893 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252070125 +0000 UTC m=+122.521402425 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-certs") pod "machine-config-server-p56l4" (UID: "0f24b5d5-6e99-4e6a-98b9-ea489c774893") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752221 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e707425-3377-458e-a331-6af00d1e989c-proxy-tls podName:1e707425-3377-458e-a331-6af00d1e989c nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252186739 +0000 UTC m=+122.521519039 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/1e707425-3377-458e-a331-6af00d1e989c-proxy-tls") pod "machine-config-operator-74547568cd-vg885" (UID: "1e707425-3377-458e-a331-6af00d1e989c") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752252 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-trusted-ca-bundle podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.25224115 +0000 UTC m=+122.521573450 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752274 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-serving-cert podName:488dbfc7-cfc6-4c83-98d4-b2bc9f16969d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252265051 +0000 UTC m=+122.521597351 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-serving-cert") pod "service-ca-operator-777779d784-znlfm" (UID: "488dbfc7-cfc6-4c83-98d4-b2bc9f16969d") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752315 4751 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752321 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-package-server-manager-serving-cert podName:82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252312873 +0000 UTC m=+122.521645173 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-xp482" (UID: "82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752414 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-session podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252389255 +0000 UTC m=+122.521721545 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-session") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752432 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-node-bootstrap-token podName:0f24b5d5-6e99-4e6a-98b9-ea489c774893 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252424436 +0000 UTC m=+122.521756716 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-node-bootstrap-token") pod "machine-config-server-p56l4" (UID: "0f24b5d5-6e99-4e6a-98b9-ea489c774893") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752447 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-serving-cert podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252438596 +0000 UTC m=+122.521770886 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752468 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-srv-cert podName:a3f6093c-dd51-45d1-9d18-0a80bac4fdb1 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252462317 +0000 UTC m=+122.521794607 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-srv-cert") pod "catalog-operator-68c6474976-rfk2f" (UID: "a3f6093c-dd51-45d1-9d18-0a80bac4fdb1") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.752889 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-service-ca podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.252872809 +0000 UTC m=+122.522205119 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753040 4751 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753176 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-srv-cert podName:e55dc627-34bc-4853-ba79-146754ae58b5 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.253163328 +0000 UTC m=+122.522495628 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-srv-cert") pod "olm-operator-6b444d44fb-fwd6f" (UID: "e55dc627-34bc-4853-ba79-146754ae58b5") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753308 4751 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753391 4751 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-ocp-branding-template: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753340 4751 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753353 4751 secret.go:188] Couldn't get secret openshift-kube-storage-version-migrator-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753563 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1e707425-3377-458e-a331-6af00d1e989c-images podName:1e707425-3377-458e-a331-6af00d1e989c nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.253398125 +0000 UTC m=+122.522730425 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/1e707425-3377-458e-a331-6af00d1e989c-images") pod "machine-config-operator-74547568cd-vg885" (UID: "1e707425-3377-458e-a331-6af00d1e989c") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753666 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-ocp-branding-template podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.253655483 +0000 UTC m=+122.522987783 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-ocp-branding-template" (UniqueName: "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-ocp-branding-template") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753731 4751 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753789 4751 secret.go:188] Couldn't get secret openshift-image-registry/image-registry-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753733 4751 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753760 4751 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-idp-0-file-data: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753764 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c2676895-4d0e-4099-9d49-abd723369982-config-volume podName:c2676895-4d0e-4099-9d49-abd723369982 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.253752596 +0000 UTC m=+122.523084896 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/c2676895-4d0e-4099-9d49-abd723369982-config-volume") pod "collect-profiles-29560320-ncm2w" (UID: "c2676895-4d0e-4099-9d49-abd723369982") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753909 4751 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753934 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9118707-2746-44ed-bc4d-004312b53b5d-serving-cert podName:c9118707-2746-44ed-bc4d-004312b53b5d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.2539036 +0000 UTC m=+122.523235930 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/c9118707-2746-44ed-bc4d-004312b53b5d-serving-cert") pod "kube-storage-version-migrator-operator-b67b599dd-c9jpp" (UID: "c9118707-2746-44ed-bc4d-004312b53b5d") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.753965 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-webhook-cert podName:8545bd4e-a993-4f46-9d64-963a5f685a2d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.253951982 +0000 UTC m=+122.523284312 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-webhook-cert") pod "packageserver-d55dfcdfc-rl4gt" (UID: "8545bd4e-a993-4f46-9d64-963a5f685a2d") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754002 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6d655c6-0506-41b7-b032-842ff9630e41-image-registry-operator-tls podName:b6d655c6-0506-41b7-b032-842ff9630e41 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.253986703 +0000 UTC m=+122.523319033 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/b6d655c6-0506-41b7-b032-842ff9630e41-image-registry-operator-tls") pod "cluster-image-registry-operator-dc59b4c8b-x9l6q" (UID: "b6d655c6-0506-41b7-b032-842ff9630e41") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754035 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-cliconfig podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254017364 +0000 UTC m=+122.523349694 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-cliconfig") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754087 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-idp-0-file-data podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254068415 +0000 UTC m=+122.523400885 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-idp-0-file-data" (UniqueName: "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-idp-0-file-data") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754130 4751 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754186 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-profile-collector-cert podName:e55dc627-34bc-4853-ba79-146754ae58b5 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254165018 +0000 UTC m=+122.523497528 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-profile-collector-cert") pod "olm-operator-6b444d44fb-fwd6f" (UID: "e55dc627-34bc-4853-ba79-146754ae58b5") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754207 4751 secret.go:188] Couldn't get secret openshift-kube-controller-manager-operator/kube-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754231 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-metrics-tls podName:bbbc093a-f5f6-4db8-89ae-1c3224e6bf75 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254209939 +0000 UTC m=+122.523542479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-metrics-tls") pod "dns-default-6d776" (UID: "bbbc093a-f5f6-4db8-89ae-1c3224e6bf75") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754270 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b60da8df-4808-44ec-8e43-bc5e7b56079d-serving-cert podName:b60da8df-4808-44ec-8e43-bc5e7b56079d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254251971 +0000 UTC m=+122.523584481 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/b60da8df-4808-44ec-8e43-bc5e7b56079d-serving-cert") pod "kube-controller-manager-operator-78b949d7b-wdvsd" (UID: "b60da8df-4808-44ec-8e43-bc5e7b56079d") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754294 4751 configmap.go:193] Couldn't get configMap openshift-multus/cni-sysctl-allowlist: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754310 4751 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754333 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9c56ee92-a6f5-46ec-8076-157268b354c7-cni-sysctl-allowlist podName:9c56ee92-a6f5-46ec-8076-157268b354c7 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254321133 +0000 UTC m=+122.523653433 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-sysctl-allowlist" (UniqueName: "kubernetes.io/configmap/9c56ee92-a6f5-46ec-8076-157268b354c7-cni-sysctl-allowlist") pod "cni-sysctl-allowlist-ds-hz7xk" (UID: "9c56ee92-a6f5-46ec-8076-157268b354c7") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754364 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/adb49d8b-968b-4fb1-8464-17853a08dda8-cert podName:adb49d8b-968b-4fb1-8464-17853a08dda8 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254347973 +0000 UTC m=+122.523680493 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/adb49d8b-968b-4fb1-8464-17853a08dda8-cert") pod "ingress-canary-nsnh2" (UID: "adb49d8b-968b-4fb1-8464-17853a08dda8") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754504 4751 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754537 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-key podName:c0f92561-e30f-4e32-80d4-b463fd87de66 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254528919 +0000 UTC m=+122.523861209 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-key") pod "service-ca-9c57cc56f-nx86z" (UID: "c0f92561-e30f-4e32-80d4-b463fd87de66") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754645 4751 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754673 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2676895-4d0e-4099-9d49-abd723369982-secret-volume podName:c2676895-4d0e-4099-9d49-abd723369982 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254665963 +0000 UTC m=+122.523998253 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/c2676895-4d0e-4099-9d49-abd723369982-secret-volume") pod "collect-profiles-29560320-ncm2w" (UID: "c2676895-4d0e-4099-9d49-abd723369982") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754703 4751 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754723 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-cabundle podName:c0f92561-e30f-4e32-80d4-b463fd87de66 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254716474 +0000 UTC m=+122.524048764 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-cabundle") pod "service-ca-9c57cc56f-nx86z" (UID: "c0f92561-e30f-4e32-80d4-b463fd87de66") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754856 4751 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.754967 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-config-volume podName:bbbc093a-f5f6-4db8-89ae-1c3224e6bf75 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.254956452 +0000 UTC m=+122.524288752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-config-volume") pod "dns-default-6d776" (UID: "bbbc093a-f5f6-4db8-89ae-1c3224e6bf75") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.755054 4751 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-login: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.755090 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-login podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.255082615 +0000 UTC m=+122.524414895 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-login" (UniqueName: "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-login") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.755376 4751 secret.go:188] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.755430 4751 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.755460 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/270222df-3862-47e2-a363-c69546415e18-proxy-tls podName:270222df-3862-47e2-a363-c69546415e18 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.255437806 +0000 UTC m=+122.524770146 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/270222df-3862-47e2-a363-c69546415e18-proxy-tls") pod "machine-config-controller-84d6567774-nh2cx" (UID: "270222df-3862-47e2-a363-c69546415e18") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.755481 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-operator-metrics podName:69dee6c1-1c2f-4d49-8955-ca4226b4306e nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.255472147 +0000 UTC m=+122.524804437 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-operator-metrics") pod "marketplace-operator-79b997595-sph8v" (UID: "69dee6c1-1c2f-4d49-8955-ca4226b4306e") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.755507 4751 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.755569 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5620fac3-59f5-49e4-9125-501353fbf836-control-plane-machine-set-operator-tls podName:5620fac3-59f5-49e4-9125-501353fbf836 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.255545649 +0000 UTC m=+122.524878189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/5620fac3-59f5-49e4-9125-501353fbf836-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-w4zmm" (UID: "5620fac3-59f5-49e4-9125-501353fbf836") : failed to sync secret cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.777824 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.783252 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.783630 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.283595149 +0000 UTC m=+122.552927479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.783974 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.785069 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.785399 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.285385673 +0000 UTC m=+122.554717973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.789634 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb16148-0952-46b4-8c72-f053bea40c0c-metrics-certs\") pod \"network-metrics-daemon-t2bcd\" (UID: \"2eb16148-0952-46b4-8c72-f053bea40c0c\") " pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.844818 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59c6efd1-70e9-45bc-8b87-94142678563d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.873286 4751 projected.go:288] Couldn't get configMap openshift-kube-scheduler-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.873352 4751 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.873435 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2d148ff3-a331-4c35-925c-3e3d0124a702-kube-api-access podName:2d148ff3-a331-4c35-925c-3e3d0124a702 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.373414948 +0000 UTC m=+122.642747238 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/2d148ff3-a331-4c35-925c-3e3d0124a702-kube-api-access") pod "openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" (UID: "2d148ff3-a331-4c35-925c-3e3d0124a702") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.874640 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t2bcd" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.875243 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.885810 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.886003 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.385956604 +0000 UTC m=+122.655288894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.886993 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.887438 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.387406547 +0000 UTC m=+122.656738847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.887621 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.907067 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.915284 4751 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.915335 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.931881 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.947745 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.981703 4751 projected.go:288] Couldn't get configMap openshift-kube-apiserver-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.982564 4751 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.982655 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-kube-api-access podName:c41e2fa1-c06c-4bdb-8a7b-11d94d47b256 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.482632428 +0000 UTC m=+122.751964718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-kube-api-access") pod "kube-apiserver-operator-766d6c64bb-74gwk" (UID: "c41e2fa1-c06c-4bdb-8a7b-11d94d47b256") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.987155 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.987914 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.988137 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.488101682 +0000 UTC m=+122.757433962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:13 crc kubenswrapper[4751]: I0316 00:08:13.988659 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:13 crc kubenswrapper[4751]: E0316 00:08:13.989113 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.489082251 +0000 UTC m=+122.758414541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.027829 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.030567 4751 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.030789 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.050984 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.061185 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-t2bcd"] Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.067579 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 16 00:08:14 crc kubenswrapper[4751]: W0316 00:08:14.070111 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2eb16148_0952_46b4_8c72_f053bea40c0c.slice/crio-a2fa04bc1d1bb4a7edaa8564cbf9cb616892fff69768c133ca779f55f7e9ac97 WatchSource:0}: Error finding container a2fa04bc1d1bb4a7edaa8564cbf9cb616892fff69768c133ca779f55f7e9ac97: Status 404 returned error can't find the container with id a2fa04bc1d1bb4a7edaa8564cbf9cb616892fff69768c133ca779f55f7e9ac97 Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.089630 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.091584 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.591562679 +0000 UTC m=+122.860894969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.140373 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vl6bg"] Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.150686 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.157830 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" event={"ID":"2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f","Type":"ContainerStarted","Data":"c881f6aad62012f265f3959a484d618a566405ce4aabb0faff175eca62a368fb"} Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.161411 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t2bcd" event={"ID":"2eb16148-0952-46b4-8c72-f053bea40c0c","Type":"ContainerStarted","Data":"a2fa04bc1d1bb4a7edaa8564cbf9cb616892fff69768c133ca779f55f7e9ac97"} Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.167475 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192473 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b695c2c9-0b76-4e93-a427-76e157f001fb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192534 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d148ff3-a331-4c35-925c-3e3d0124a702-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192564 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-metrics-certs\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192583 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-console-config\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192614 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192721 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-serving-cert\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192755 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-config\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192773 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-client-ca\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192802 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.192891 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bzhm\" (UniqueName: \"kubernetes.io/projected/da8ac6ff-99fb-4c50-a706-eec6c8987b04-kube-api-access-8bzhm\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.193253 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8kxm\" (UniqueName: \"kubernetes.io/projected/7e4947be-bb74-46c6-9a89-9c44dede6c7c-kube-api-access-l8kxm\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.193396 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-oauth-config\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.195710 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.695685656 +0000 UTC m=+122.965017946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.195954 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfe3c553-b411-4074-b83a-c8eba2518089-console-config\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.196709 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74e4f428-25a3-45b9-959e-8a1eeb968cdf-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.197475 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-config\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.198050 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-client-ca\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.198843 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-serving-cert\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.198918 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d148ff3-a331-4c35-925c-3e3d0124a702-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.199992 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b695c2c9-0b76-4e93-a427-76e157f001fb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.203747 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bzhm\" (UniqueName: \"kubernetes.io/projected/da8ac6ff-99fb-4c50-a706-eec6c8987b04-kube-api-access-8bzhm\") pod \"console-operator-58897d9998-njjzb\" (UID: \"da8ac6ff-99fb-4c50-a706-eec6c8987b04\") " pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.203969 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a792c668-ee31-4bd8-a122-37bf8b48f37f-metrics-certs\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.204710 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8kxm\" (UniqueName: \"kubernetes.io/projected/7e4947be-bb74-46c6-9a89-9c44dede6c7c-kube-api-access-l8kxm\") pod \"etcd-operator-b45778765-jszhl\" (UID: \"7e4947be-bb74-46c6-9a89-9c44dede6c7c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.204911 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfe3c553-b411-4074-b83a-c8eba2518089-console-oauth-config\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.205005 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b6d655c6-0506-41b7-b032-842ff9630e41-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.206121 4751 projected.go:288] Couldn't get configMap openshift-machine-api/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.229164 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk"] Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.234406 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.261258 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcpx7\" (UniqueName: \"kubernetes.io/projected/0f24b5d5-6e99-4e6a-98b9-ea489c774893-kube-api-access-rcpx7\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.294263 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.294488 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.794465013 +0000 UTC m=+123.063797303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295360 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295416 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-srv-cert\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295454 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9118707-2746-44ed-bc4d-004312b53b5d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295490 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2676895-4d0e-4099-9d49-abd723369982-config-volume\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295523 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1e707425-3377-458e-a331-6af00d1e989c-images\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295586 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/adb49d8b-968b-4fb1-8464-17853a08dda8-cert\") pod \"ingress-canary-nsnh2\" (UID: \"adb49d8b-968b-4fb1-8464-17853a08dda8\") " pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295650 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-key\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295708 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-metrics-tls\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295730 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2676895-4d0e-4099-9d49-abd723369982-secret-volume\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295836 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-cabundle\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295866 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295898 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5620fac3-59f5-49e4-9125-501353fbf836-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-w4zmm\" (UID: \"5620fac3-59f5-49e4-9125-501353fbf836\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295925 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/270222df-3862-47e2-a363-c69546415e18-proxy-tls\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295948 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295977 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-apiservice-cert\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.295998 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296018 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296041 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-config\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296070 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296096 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9118707-2746-44ed-bc4d-004312b53b5d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296157 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-policies\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296214 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296236 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b60da8df-4808-44ec-8e43-bc5e7b56079d-config\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296256 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xp482\" (UID: \"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296284 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296316 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296340 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-profile-collector-cert\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296362 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-serving-cert\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296393 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296419 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1ee3303-f745-4058-8927-046b07740c33-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6qwkx\" (UID: \"c1ee3303-f745-4058-8927-046b07740c33\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296440 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296464 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-webhook-cert\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296480 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296519 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6d655c6-0506-41b7-b032-842ff9630e41-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296540 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296568 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-srv-cert\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296585 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60da8df-4808-44ec-8e43-bc5e7b56079d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296627 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9c56ee92-a6f5-46ec-8076-157268b354c7-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296654 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296965 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-certs\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.296993 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-node-bootstrap-token\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.297012 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.297040 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-config-volume\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.297090 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e707425-3377-458e-a331-6af00d1e989c-proxy-tls\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.297492 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.797479994 +0000 UTC m=+123.066812274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.298284 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9118707-2746-44ed-bc4d-004312b53b5d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.299290 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-policies\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.299590 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-config\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.301014 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.301203 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-apiservice-cert\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.303694 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vn22\" (UniqueName: \"kubernetes.io/projected/b6d655c6-0506-41b7-b032-842ff9630e41-kube-api-access-5vn22\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.304292 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.304497 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8545bd4e-a993-4f46-9d64-963a5f685a2d-webhook-cert\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.307171 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.314323 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.328750 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.341760 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xp482\" (UID: \"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.346933 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.353278 4751 projected.go:288] Couldn't get configMap openshift-console/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.358528 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b60da8df-4808-44ec-8e43-bc5e7b56079d-config\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.366689 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.369884 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.387271 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.397392 4751 projected.go:288] Couldn't get configMap openshift-ingress/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.397571 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.397693 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.897676733 +0000 UTC m=+123.167009023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.397984 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.398027 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d148ff3-a331-4c35-925c-3e3d0124a702-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.398573 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.89855876 +0000 UTC m=+123.167891050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.400918 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.402095 4751 projected.go:288] Couldn't get configMap openshift-cluster-machine-approver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.406898 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.410473 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-profile-collector-cert\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.411428 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.419513 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2676895-4d0e-4099-9d49-abd723369982-secret-volume\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.432600 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.440778 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-serving-cert\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.454589 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.461545 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.467381 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.474182 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1ee3303-f745-4058-8927-046b07740c33-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6qwkx\" (UID: \"c1ee3303-f745-4058-8927-046b07740c33\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.494026 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.497718 4751 projected.go:288] Couldn't get configMap openshift-config-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.498559 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.499057 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.499344 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:14.999316176 +0000 UTC m=+123.268648486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.499726 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.499866 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.500095 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.000082519 +0000 UTC m=+123.269414809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.507126 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.513989 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-certs\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.526155 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.536210 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0f24b5d5-6e99-4e6a-98b9-ea489c774893-node-bootstrap-token\") pod \"machine-config-server-p56l4\" (UID: \"0f24b5d5-6e99-4e6a-98b9-ea489c774893\") " pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.560720 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d8ht\" (UniqueName: \"kubernetes.io/projected/9c56ee92-a6f5-46ec-8076-157268b354c7-kube-api-access-8d8ht\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.568854 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.572079 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.600971 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.601228 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.101208997 +0000 UTC m=+123.370541287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.601821 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.602191 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.102183186 +0000 UTC m=+123.371515476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.603649 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsfc2\" (UniqueName: \"kubernetes.io/projected/1e707425-3377-458e-a331-6af00d1e989c-kube-api-access-gsfc2\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.604794 4751 projected.go:288] Couldn't get configMap openshift-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.626263 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv9n2\" (UniqueName: \"kubernetes.io/projected/270222df-3862-47e2-a363-c69546415e18-kube-api-access-vv9n2\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.646017 4751 request.go:700] Waited for 1.892648951s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/serviceaccounts/csi-hostpath-provisioner-sa/token Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.646032 4751 projected.go:288] Couldn't get configMap openshift-console/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.690151 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.691517 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.701796 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e55dc627-34bc-4853-ba79-146754ae58b5-srv-cert\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.702507 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.702715 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.202684635 +0000 UTC m=+123.472016935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.703590 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.703696 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.704041 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.204030915 +0000 UTC m=+123.473363215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.705950 4751 projected.go:288] Couldn't get configMap openshift-controller-manager-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.707184 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.710987 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9118707-2746-44ed-bc4d-004312b53b5d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.729357 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.740293 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2676895-4d0e-4099-9d49-abd723369982-config-volume\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.747302 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.749003 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1e707425-3377-458e-a331-6af00d1e989c-images\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.787517 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.802698 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e707425-3377-458e-a331-6af00d1e989c-proxy-tls\") pod \"machine-config-operator-74547568cd-vg885\" (UID: \"1e707425-3377-458e-a331-6af00d1e989c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.804294 4751 projected.go:288] Couldn't get configMap openshift-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.804771 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.804946 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.304923866 +0000 UTC m=+123.574256166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.805181 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.805559 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.305550055 +0000 UTC m=+123.574882355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.807076 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.811569 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-config-volume\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.828223 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.829287 4751 projected.go:288] Couldn't get configMap openshift-cluster-samples-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.838601 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.846785 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.855557 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.864446 4751 projected.go:288] Couldn't get configMap openshift-ingress-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.869397 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.877337 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6d655c6-0506-41b7-b032-842ff9630e41-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-x9l6q\" (UID: \"b6d655c6-0506-41b7-b032-842ff9630e41\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.906641 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.906859 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.406824727 +0000 UTC m=+123.676157027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.907680 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.908191 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 16 00:08:14 crc kubenswrapper[4751]: E0316 00:08:14.908260 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.408241259 +0000 UTC m=+123.677573589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.922504 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-metrics-tls\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.947540 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.955497 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60da8df-4808-44ec-8e43-bc5e7b56079d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.966999 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.970092 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9c56ee92-a6f5-46ec-8076-157268b354c7-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-hz7xk\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.987287 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 16 00:08:14 crc kubenswrapper[4751]: I0316 00:08:14.993469 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-srv-cert\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.008380 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.008571 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.508527731 +0000 UTC m=+123.777860061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.009158 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.009619 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.509597673 +0000 UTC m=+123.778930003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.013250 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.020825 4751 projected.go:288] Couldn't get configMap openshift-kube-controller-manager-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.020873 4751 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.020944 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b60da8df-4808-44ec-8e43-bc5e7b56079d-kube-api-access podName:b60da8df-4808-44ec-8e43-bc5e7b56079d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.520924342 +0000 UTC m=+123.790256672 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/b60da8df-4808-44ec-8e43-bc5e7b56079d-kube-api-access") pod "kube-controller-manager-operator-78b949d7b-wdvsd" (UID: "b60da8df-4808-44ec-8e43-bc5e7b56079d") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.024425 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/adb49d8b-968b-4fb1-8464-17853a08dda8-cert\") pod \"ingress-canary-nsnh2\" (UID: \"adb49d8b-968b-4fb1-8464-17853a08dda8\") " pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.067956 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.078267 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8bzf\" (UniqueName: \"kubernetes.io/projected/bbbc093a-f5f6-4db8-89ae-1c3224e6bf75-kube-api-access-h8bzf\") pod \"dns-default-6d776\" (UID: \"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75\") " pod="openshift-dns/dns-default-6d776" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.082277 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-key\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.087535 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.089150 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c0f92561-e30f-4e32-80d4-b463fd87de66-signing-cabundle\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.101501 4751 projected.go:288] Couldn't get configMap openshift-operator-lifecycle-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.110827 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.110961 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.610932567 +0000 UTC m=+123.880264887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.111254 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.111699 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.61168236 +0000 UTC m=+123.881014680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.130326 4751 projected.go:288] Couldn't get configMap openshift-operator-lifecycle-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.136416 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.141615 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.142774 4751 projected.go:288] Couldn't get configMap openshift-ingress-canary/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.154208 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.167859 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.206734 4751 projected.go:288] Couldn't get configMap openshift-machine-api/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.206780 4751 projected.go:194] Error preparing data for projected volume kube-api-access-dj7k6 for pod openshift-machine-api/machine-api-operator-5694c8668f-zb5z5: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.206910 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a932e9e3-ca99-4af3-a9e6-8888e31be60d-kube-api-access-dj7k6 podName:a932e9e3-ca99-4af3-a9e6-8888e31be60d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.70686829 +0000 UTC m=+123.976200630 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-dj7k6" (UniqueName: "kubernetes.io/projected/a932e9e3-ca99-4af3-a9e6-8888e31be60d-kube-api-access-dj7k6") pod "machine-api-operator-5694c8668f-zb5z5" (UID: "a932e9e3-ca99-4af3-a9e6-8888e31be60d") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.208413 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/270222df-3862-47e2-a363-c69546415e18-proxy-tls\") pod \"machine-config-controller-84d6567774-nh2cx\" (UID: \"270222df-3862-47e2-a363-c69546415e18\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.208615 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.212213 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.212392 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.712368844 +0000 UTC m=+123.981701164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.212828 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.213364 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.713349904 +0000 UTC m=+123.982682304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.214019 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t2bcd" event={"ID":"2eb16148-0952-46b4-8c72-f053bea40c0c","Type":"ContainerStarted","Data":"158d6273a2144665aac6ddd7a10925ad9e1e2c69f13021091416622f6a490abc"} Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.214088 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t2bcd" event={"ID":"2eb16148-0952-46b4-8c72-f053bea40c0c","Type":"ContainerStarted","Data":"d642e2ab8725bd9b8bf4d6b66af30bcac7b6e4022de167dc57dc18b3cd5621f5"} Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.214334 4751 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-console-operator/console-operator-58897d9998-njjzb" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.214421 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.216073 4751 generic.go:334] "Generic (PLEG): container finished" podID="8e87f4be-4c72-4d7d-a2b7-4584d11be3c8" containerID="e8765a36d832aa1c3b124e7a37eafb452c527138dae4fc3b5bb37058cab6f2ad" exitCode=0 Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.216237 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" event={"ID":"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8","Type":"ContainerDied","Data":"e8765a36d832aa1c3b124e7a37eafb452c527138dae4fc3b5bb37058cab6f2ad"} Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.216270 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" event={"ID":"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8","Type":"ContainerStarted","Data":"984e4e0a59b91cc08ced014d6d8f3a0b868f47470d0c1cc3ac6713daade2ba86"} Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.217285 4751 projected.go:288] Couldn't get configMap openshift-authentication/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.217345 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.219132 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" event={"ID":"2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f","Type":"ContainerStarted","Data":"5721b2a974b63c1a538e0e8becbb209731d5a9abdf677fd4d229d895bb988552"} Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.219231 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" event={"ID":"2b2f6e3c-e3bd-4dd8-b7d8-3106cfa45e6f","Type":"ContainerStarted","Data":"40a8b65d90e22e1059a711d49e0d67eb606d6efee731f43efaab4505b2b34d30"} Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.241433 4751 projected.go:288] Couldn't get configMap openshift-service-ca-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.244977 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj5tl\" (UniqueName: \"kubernetes.io/projected/c1ee3303-f745-4058-8927-046b07740c33-kube-api-access-vj5tl\") pod \"multus-admission-controller-857f4d67dd-6qwkx\" (UID: \"c1ee3303-f745-4058-8927-046b07740c33\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.247053 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.247082 4751 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.247346 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.255357 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5620fac3-59f5-49e4-9125-501353fbf836-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-w4zmm\" (UID: \"5620fac3-59f5-49e4-9125-501353fbf836\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.271345 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.282819 4751 projected.go:288] Couldn't get configMap openshift-kube-storage-version-migrator-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.287749 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.289908 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d148ff3-a331-4c35-925c-3e3d0124a702-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr4v6\" (UID: \"2d148ff3-a331-4c35-925c-3e3d0124a702\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.308333 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.314736 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.316848 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.816805331 +0000 UTC m=+124.086137681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.317734 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.318633 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.818618455 +0000 UTC m=+124.087950895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.317360 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c41e2fa1-c06c-4bdb-8a7b-11d94d47b256-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-74gwk\" (UID: \"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.331631 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.350352 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.359408 4751 projected.go:288] Couldn't get configMap openshift-console/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.359468 4751 projected.go:194] Error preparing data for projected volume kube-api-access-6ngn4 for pod openshift-console/console-f9d7485db-bd5b9: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.359557 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfe3c553-b411-4074-b83a-c8eba2518089-kube-api-access-6ngn4 podName:cfe3c553-b411-4074-b83a-c8eba2518089 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.85952944 +0000 UTC m=+124.128861740 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6ngn4" (UniqueName: "kubernetes.io/projected/cfe3c553-b411-4074-b83a-c8eba2518089-kube-api-access-6ngn4") pod "console-f9d7485db-bd5b9" (UID: "cfe3c553-b411-4074-b83a-c8eba2518089") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.366728 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.386646 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.401411 4751 projected.go:288] Couldn't get configMap openshift-ingress/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.401449 4751 projected.go:194] Error preparing data for projected volume kube-api-access-nfgxp for pod openshift-ingress/router-default-5444994796-jpvk6: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.401532 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a792c668-ee31-4bd8-a122-37bf8b48f37f-kube-api-access-nfgxp podName:a792c668-ee31-4bd8-a122-37bf8b48f37f nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.901509196 +0000 UTC m=+124.170841496 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-nfgxp" (UniqueName: "kubernetes.io/projected/a792c668-ee31-4bd8-a122-37bf8b48f37f-kube-api-access-nfgxp") pod "router-default-5444994796-jpvk6" (UID: "a792c668-ee31-4bd8-a122-37bf8b48f37f") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.402239 4751 projected.go:288] Couldn't get configMap openshift-cluster-machine-approver/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.402278 4751 projected.go:194] Error preparing data for projected volume kube-api-access-l2fgg for pod openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.402363 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-kube-api-access-l2fgg podName:122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.902341181 +0000 UTC m=+124.171673481 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-l2fgg" (UniqueName: "kubernetes.io/projected/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-kube-api-access-l2fgg") pod "machine-approver-56656f9798-jlhhf" (UID: "122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.409885 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.421182 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.421351 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.92132381 +0000 UTC m=+124.190656110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.421691 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.422004 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.92198978 +0000 UTC m=+124.191322070 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.427652 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.447089 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.458358 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-njjzb"] Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.466876 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: W0316 00:08:15.467628 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda8ac6ff_99fb_4c50_a706_eec6c8987b04.slice/crio-0d188dedcb196f2f42157b5fa69c9d1550faedb03376136857899666daf075e8 WatchSource:0}: Error finding container 0d188dedcb196f2f42157b5fa69c9d1550faedb03376136857899666daf075e8: Status 404 returned error can't find the container with id 0d188dedcb196f2f42157b5fa69c9d1550faedb03376136857899666daf075e8 Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.487283 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.498066 4751 projected.go:288] Couldn't get configMap openshift-config-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.498087 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cdgj4 for pod openshift-config-operator/openshift-config-operator-7777fb866f-88ldb: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.498164 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-kube-api-access-cdgj4 podName:3b5183fc-18b0-4aa5-9d1c-de26657da9f7 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:15.99814506 +0000 UTC m=+124.267477350 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cdgj4" (UniqueName: "kubernetes.io/projected/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-kube-api-access-cdgj4") pod "openshift-config-operator-7777fb866f-88ldb" (UID: "3b5183fc-18b0-4aa5-9d1c-de26657da9f7") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.502314 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jszhl"] Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.506720 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: W0316 00:08:15.510965 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e4947be_bb74_46c6_9a89_9c44dede6c7c.slice/crio-c948c69628031685b89fcc80f8a23b63418a37ea33ed9e99c2882b2f8ebe0bc7 WatchSource:0}: Error finding container c948c69628031685b89fcc80f8a23b63418a37ea33ed9e99c2882b2f8ebe0bc7: Status 404 returned error can't find the container with id c948c69628031685b89fcc80f8a23b63418a37ea33ed9e99c2882b2f8ebe0bc7 Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.522264 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.522466 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.022431717 +0000 UTC m=+124.291764007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.522784 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60da8df-4808-44ec-8e43-bc5e7b56079d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.522814 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.523273 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.023258091 +0000 UTC m=+124.292590381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.527601 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.567009 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.577677 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60da8df-4808-44ec-8e43-bc5e7b56079d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wdvsd\" (UID: \"b60da8df-4808-44ec-8e43-bc5e7b56079d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.588983 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.605996 4751 projected.go:288] Couldn't get configMap openshift-controller-manager/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.606034 4751 projected.go:194] Error preparing data for projected volume kube-api-access-hnqqr for pod openshift-controller-manager/controller-manager-879f6c89f-mrm4x: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.606091 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a8fdc932-6670-45f5-9f69-c15eaae40bc4-kube-api-access-hnqqr podName:a8fdc932-6670-45f5-9f69-c15eaae40bc4 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.106073411 +0000 UTC m=+124.375405701 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hnqqr" (UniqueName: "kubernetes.io/projected/a8fdc932-6670-45f5-9f69-c15eaae40bc4-kube-api-access-hnqqr") pod "controller-manager-879f6c89f-mrm4x" (UID: "a8fdc932-6670-45f5-9f69-c15eaae40bc4") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.607515 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.623383 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.623601 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.123566824 +0000 UTC m=+124.392899124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.623952 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.624324 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.124310217 +0000 UTC m=+124.393642507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.626461 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.633203 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksrcg\" (UniqueName: \"kubernetes.io/projected/5620fac3-59f5-49e4-9125-501353fbf836-kube-api-access-ksrcg\") pod \"control-plane-machine-set-operator-78cbb6b69f-w4zmm\" (UID: \"5620fac3-59f5-49e4-9125-501353fbf836\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.639846 4751 projected.go:288] Couldn't get configMap openshift-marketplace/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.647077 4751 projected.go:288] Couldn't get configMap openshift-console/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.647128 4751 projected.go:194] Error preparing data for projected volume kube-api-access-tw4s5 for pod openshift-console/downloads-7954f5f757-6k5cn: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.647238 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2394c21b-2d69-4698-8e58-d0198ec77558-kube-api-access-tw4s5 podName:2394c21b-2d69-4698-8e58-d0198ec77558 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.147216512 +0000 UTC m=+124.416548892 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-tw4s5" (UniqueName: "kubernetes.io/projected/2394c21b-2d69-4698-8e58-d0198ec77558-kube-api-access-tw4s5") pod "downloads-7954f5f757-6k5cn" (UID: "2394c21b-2d69-4698-8e58-d0198ec77558") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.648209 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.658806 4751 projected.go:288] Couldn't get configMap hostpath-provisioner/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.665613 4751 request.go:700] Waited for 1.448227847s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27100 Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.667606 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.686929 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.707223 4751 projected.go:288] Couldn't get configMap openshift-controller-manager-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.707259 4751 projected.go:194] Error preparing data for projected volume kube-api-access-xwvwk for pod openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.707390 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-kube-api-access-xwvwk podName:a230e2c9-cb14-4f26-8552-d81e42a7fcc2 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.207364403 +0000 UTC m=+124.476696703 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-xwvwk" (UniqueName: "kubernetes.io/projected/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-kube-api-access-xwvwk") pod "openshift-controller-manager-operator-756b6f6bc6-d528n" (UID: "a230e2c9-cb14-4f26-8552-d81e42a7fcc2") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.709162 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.724750 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.724826 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.224810765 +0000 UTC m=+124.494143055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.725038 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj7k6\" (UniqueName: \"kubernetes.io/projected/a932e9e3-ca99-4af3-a9e6-8888e31be60d-kube-api-access-dj7k6\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.725266 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.726161 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.226152576 +0000 UTC m=+124.495484856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.731649 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.731781 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj7k6\" (UniqueName: \"kubernetes.io/projected/a932e9e3-ca99-4af3-a9e6-8888e31be60d-kube-api-access-dj7k6\") pod \"machine-api-operator-5694c8668f-zb5z5\" (UID: \"a932e9e3-ca99-4af3-a9e6-8888e31be60d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.748732 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.767740 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.787960 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.804791 4751 projected.go:288] Couldn't get configMap openshift-apiserver/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.804837 4751 projected.go:194] Error preparing data for projected volume kube-api-access-nvr6b for pod openshift-apiserver/apiserver-76f77b778f-9mrng: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.804919 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/74e4f428-25a3-45b9-959e-8a1eeb968cdf-kube-api-access-nvr6b podName:74e4f428-25a3-45b9-959e-8a1eeb968cdf nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.304897293 +0000 UTC m=+124.574229583 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-nvr6b" (UniqueName: "kubernetes.io/projected/74e4f428-25a3-45b9-959e-8a1eeb968cdf-kube-api-access-nvr6b") pod "apiserver-76f77b778f-9mrng" (UID: "74e4f428-25a3-45b9-959e-8a1eeb968cdf") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.807769 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.813298 4751 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-machine-config-operator/machine-config-server-p56l4" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.813388 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-p56l4" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.826648 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.827316 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.327291514 +0000 UTC m=+124.596623824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.828675 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.829914 4751 projected.go:288] Couldn't get configMap openshift-cluster-samples-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.829948 4751 projected.go:194] Error preparing data for projected volume kube-api-access-jk95r for pod openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.830020 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1c921c65-0fb5-4d15-b303-c9b7d0021931-kube-api-access-jk95r podName:1c921c65-0fb5-4d15-b303-c9b7d0021931 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.329984264 +0000 UTC m=+124.599316574 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-jk95r" (UniqueName: "kubernetes.io/projected/1c921c65-0fb5-4d15-b303-c9b7d0021931-kube-api-access-jk95r") pod "cluster-samples-operator-665b6dd947-qsvvx" (UID: "1c921c65-0fb5-4d15-b303-c9b7d0021931") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.848179 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.865051 4751 projected.go:288] Couldn't get configMap openshift-ingress-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.865083 4751 projected.go:194] Error preparing data for projected volume kube-api-access-84j5b for pod openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.865172 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/59c6efd1-70e9-45bc-8b87-94142678563d-kube-api-access-84j5b podName:59c6efd1-70e9-45bc-8b87-94142678563d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.365149517 +0000 UTC m=+124.634481807 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-84j5b" (UniqueName: "kubernetes.io/projected/59c6efd1-70e9-45bc-8b87-94142678563d-kube-api-access-84j5b") pod "ingress-operator-5b745b69d9-k5qzx" (UID: "59c6efd1-70e9-45bc-8b87-94142678563d") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.866632 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.887205 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.907732 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.918626 4751 projected.go:288] Couldn't get configMap openshift-service-ca/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.927900 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ngn4\" (UniqueName: \"kubernetes.io/projected/cfe3c553-b411-4074-b83a-c8eba2518089-kube-api-access-6ngn4\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.927941 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfgxp\" (UniqueName: \"kubernetes.io/projected/a792c668-ee31-4bd8-a122-37bf8b48f37f-kube-api-access-nfgxp\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.927967 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2fgg\" (UniqueName: \"kubernetes.io/projected/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-kube-api-access-l2fgg\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.928116 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.928191 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 16 00:08:15 crc kubenswrapper[4751]: E0316 00:08:15.929435 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.42940024 +0000 UTC m=+124.698732640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.932916 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ngn4\" (UniqueName: \"kubernetes.io/projected/cfe3c553-b411-4074-b83a-c8eba2518089-kube-api-access-6ngn4\") pod \"console-f9d7485db-bd5b9\" (UID: \"cfe3c553-b411-4074-b83a-c8eba2518089\") " pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.933738 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2fgg\" (UniqueName: \"kubernetes.io/projected/122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa-kube-api-access-l2fgg\") pod \"machine-approver-56656f9798-jlhhf\" (UID: \"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.935463 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfgxp\" (UniqueName: \"kubernetes.io/projected/a792c668-ee31-4bd8-a122-37bf8b48f37f-kube-api-access-nfgxp\") pod \"router-default-5444994796-jpvk6\" (UID: \"a792c668-ee31-4bd8-a122-37bf8b48f37f\") " pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.947910 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.960877 4751 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.960917 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.971063 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 16 00:08:15 crc kubenswrapper[4751]: I0316 00:08:15.988033 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.000932 4751 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.001046 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.006812 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.026874 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.029832 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.030175 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdgj4\" (UniqueName: \"kubernetes.io/projected/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-kube-api-access-cdgj4\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.030390 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.530363143 +0000 UTC m=+124.799695453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.034809 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdgj4\" (UniqueName: \"kubernetes.io/projected/3b5183fc-18b0-4aa5-9d1c-de26657da9f7-kube-api-access-cdgj4\") pod \"openshift-config-operator-7777fb866f-88ldb\" (UID: \"3b5183fc-18b0-4aa5-9d1c-de26657da9f7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.046082 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.066945 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.070879 4751 projected.go:194] Error preparing data for projected volume kube-api-access-bh7cp for pod openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.071022 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-kube-api-access-bh7cp podName:82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.570988229 +0000 UTC m=+124.840320539 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-bh7cp" (UniqueName: "kubernetes.io/projected/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-kube-api-access-bh7cp") pod "package-server-manager-789f6589d5-xp482" (UID: "82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.071978 4751 projected.go:194] Error preparing data for projected volume kube-api-access-m8lhg for pod openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.072026 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8545bd4e-a993-4f46-9d64-963a5f685a2d-kube-api-access-m8lhg podName:8545bd4e-a993-4f46-9d64-963a5f685a2d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.57201418 +0000 UTC m=+124.841346480 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-m8lhg" (UniqueName: "kubernetes.io/projected/8545bd4e-a993-4f46-9d64-963a5f685a2d-kube-api-access-m8lhg") pod "packageserver-d55dfcdfc-rl4gt" (UID: "8545bd4e-a993-4f46-9d64-963a5f685a2d") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.079769 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfhrv\" (UniqueName: \"kubernetes.io/projected/e55dc627-34bc-4853-ba79-146754ae58b5-kube-api-access-mfhrv\") pod \"olm-operator-6b444d44fb-fwd6f\" (UID: \"e55dc627-34bc-4853-ba79-146754ae58b5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.080164 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79vjv\" (UniqueName: \"kubernetes.io/projected/a3f6093c-dd51-45d1-9d18-0a80bac4fdb1-kube-api-access-79vjv\") pod \"catalog-operator-68c6474976-rfk2f\" (UID: \"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.081147 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76wnr\" (UniqueName: \"kubernetes.io/projected/c2676895-4d0e-4099-9d49-abd723369982-kube-api-access-76wnr\") pod \"collect-profiles-29560320-ncm2w\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.090043 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.093816 4751 projected.go:194] Error preparing data for projected volume kube-api-access-mmdpq for pod openshift-ingress-canary/ingress-canary-nsnh2: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.093877 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/adb49d8b-968b-4fb1-8464-17853a08dda8-kube-api-access-mmdpq podName:adb49d8b-968b-4fb1-8464-17853a08dda8 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.593859394 +0000 UTC m=+124.863191684 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mmdpq" (UniqueName: "kubernetes.io/projected/adb49d8b-968b-4fb1-8464-17853a08dda8-kube-api-access-mmdpq") pod "ingress-canary-nsnh2" (UID: "adb49d8b-968b-4fb1-8464-17853a08dda8") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.112924 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.128694 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.135079 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnqqr\" (UniqueName: \"kubernetes.io/projected/a8fdc932-6670-45f5-9f69-c15eaae40bc4-kube-api-access-hnqqr\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.135218 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6d776" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.135312 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.135883 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.635862122 +0000 UTC m=+124.905194422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.140533 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnqqr\" (UniqueName: \"kubernetes.io/projected/a8fdc932-6670-45f5-9f69-c15eaae40bc4-kube-api-access-hnqqr\") pod \"controller-manager-879f6c89f-mrm4x\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.147141 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.148382 4751 projected.go:194] Error preparing data for projected volume kube-api-access-shm2f for pod openshift-authentication/oauth-openshift-558db77b4-sfnlp: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.148668 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-kube-api-access-shm2f podName:5f11ca1f-26b5-4887-92c5-cdbf4dac90ec nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.648465359 +0000 UTC m=+124.917797649 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-shm2f" (UniqueName: "kubernetes.io/projected/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-kube-api-access-shm2f") pod "oauth-openshift-558db77b4-sfnlp" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.167442 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.171669 4751 projected.go:194] Error preparing data for projected volume kube-api-access-69gt4 for pod openshift-service-ca-operator/service-ca-operator-777779d784-znlfm: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.171763 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-kube-api-access-69gt4 podName:488dbfc7-cfc6-4c83-98d4-b2bc9f16969d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.671737906 +0000 UTC m=+124.941070216 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-69gt4" (UniqueName: "kubernetes.io/projected/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-kube-api-access-69gt4") pod "service-ca-operator-777779d784-znlfm" (UID: "488dbfc7-cfc6-4c83-98d4-b2bc9f16969d") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.187472 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.194864 4751 projected.go:194] Error preparing data for projected volume kube-api-access-pb49d for pod openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.194942 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c9118707-2746-44ed-bc4d-004312b53b5d-kube-api-access-pb49d podName:c9118707-2746-44ed-bc4d-004312b53b5d nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.6949206 +0000 UTC m=+124.964252900 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-pb49d" (UniqueName: "kubernetes.io/projected/c9118707-2746-44ed-bc4d-004312b53b5d-kube-api-access-pb49d") pod "kube-storage-version-migrator-operator-b67b599dd-c9jpp" (UID: "c9118707-2746-44ed-bc4d-004312b53b5d") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.199402 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q"] Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.206609 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.215220 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.233580 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.234996 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" event={"ID":"8e87f4be-4c72-4d7d-a2b7-4584d11be3c8","Type":"ContainerStarted","Data":"a07bc34fbd421602291f791e7919b0b0afb34d60d9c6249b7c6fb679d91a8437"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.235961 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.236340 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw4s5\" (UniqueName: \"kubernetes.io/projected/2394c21b-2d69-4698-8e58-d0198ec77558-kube-api-access-tw4s5\") pod \"downloads-7954f5f757-6k5cn\" (UID: \"2394c21b-2d69-4698-8e58-d0198ec77558\") " pod="openshift-console/downloads-7954f5f757-6k5cn" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.236368 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwvwk\" (UniqueName: \"kubernetes.io/projected/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-kube-api-access-xwvwk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.236886 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.736863985 +0000 UTC m=+125.006196275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.237523 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" event={"ID":"7e4947be-bb74-46c6-9a89-9c44dede6c7c","Type":"ContainerStarted","Data":"e1655c1f62d588d9364925f1bb866e0a61760b93c028531f4d562a50475222d9"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.237579 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" event={"ID":"7e4947be-bb74-46c6-9a89-9c44dede6c7c","Type":"ContainerStarted","Data":"c948c69628031685b89fcc80f8a23b63418a37ea33ed9e99c2882b2f8ebe0bc7"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.239077 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vg885"] Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.239193 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.239534 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwvwk\" (UniqueName: \"kubernetes.io/projected/a230e2c9-cb14-4f26-8552-d81e42a7fcc2-kube-api-access-xwvwk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d528n\" (UID: \"a230e2c9-cb14-4f26-8552-d81e42a7fcc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.244340 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw4s5\" (UniqueName: \"kubernetes.io/projected/2394c21b-2d69-4698-8e58-d0198ec77558-kube-api-access-tw4s5\") pod \"downloads-7954f5f757-6k5cn\" (UID: \"2394c21b-2d69-4698-8e58-d0198ec77558\") " pod="openshift-console/downloads-7954f5f757-6k5cn" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.244724 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-p56l4" event={"ID":"0f24b5d5-6e99-4e6a-98b9-ea489c774893","Type":"ContainerStarted","Data":"9698d98a64e882057760aeb7b6550661a21f6e9701adc468249bcd1ce9458892"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.244767 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-p56l4" event={"ID":"0f24b5d5-6e99-4e6a-98b9-ea489c774893","Type":"ContainerStarted","Data":"d2313dd3caccda04d01919b50f1dba42a1c521020968fee0e5a251bef60626e0"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.247402 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.254441 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" event={"ID":"9c56ee92-a6f5-46ec-8076-157268b354c7","Type":"ContainerStarted","Data":"80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.254506 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" event={"ID":"9c56ee92-a6f5-46ec-8076-157268b354c7","Type":"ContainerStarted","Data":"10ba323e9d8c51e682d7c9ed58e560dcf8554ed1a0a6bc5e06bd607d18b50c79"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.254760 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.262801 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.291731 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.292521 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.306678 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.307780 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.310888 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-njjzb" event={"ID":"da8ac6ff-99fb-4c50-a706-eec6c8987b04","Type":"ContainerStarted","Data":"012d5df77f699ff56fe2035f263fa9904d7f00765c8f2219538fa0034f64a152"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.310928 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-njjzb" event={"ID":"da8ac6ff-99fb-4c50-a706-eec6c8987b04","Type":"ContainerStarted","Data":"0d188dedcb196f2f42157b5fa69c9d1550faedb03376136857899666daf075e8"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.312116 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.316050 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.326627 4751 patch_prober.go:28] interesting pod/console-operator-58897d9998-njjzb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.326677 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-njjzb" podUID="da8ac6ff-99fb-4c50-a706-eec6c8987b04" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.327275 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.332205 4751 projected.go:194] Error preparing data for projected volume kube-api-access-fk6sg for pod openshift-marketplace/marketplace-operator-79b997595-sph8v: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.332274 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69dee6c1-1c2f-4d49-8955-ca4226b4306e-kube-api-access-fk6sg podName:69dee6c1-1c2f-4d49-8955-ca4226b4306e nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.832256121 +0000 UTC m=+125.101588411 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-fk6sg" (UniqueName: "kubernetes.io/projected/69dee6c1-1c2f-4d49-8955-ca4226b4306e-kube-api-access-fk6sg") pod "marketplace-operator-79b997595-sph8v" (UID: "69dee6c1-1c2f-4d49-8955-ca4226b4306e") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.337496 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk95r\" (UniqueName: \"kubernetes.io/projected/1c921c65-0fb5-4d15-b303-c9b7d0021931-kube-api-access-jk95r\") pod \"cluster-samples-operator-665b6dd947-qsvvx\" (UID: \"1c921c65-0fb5-4d15-b303-c9b7d0021931\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.337528 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvr6b\" (UniqueName: \"kubernetes.io/projected/74e4f428-25a3-45b9-959e-8a1eeb968cdf-kube-api-access-nvr6b\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.337574 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.339031 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.839009523 +0000 UTC m=+125.108341903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.339202 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" event={"ID":"b6d655c6-0506-41b7-b032-842ff9630e41","Type":"ContainerStarted","Data":"672aae02d6b0204f63f640c24f1cd15d0a3aa9ce41b88c081486d2cb8a609004"} Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.345696 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk95r\" (UniqueName: \"kubernetes.io/projected/1c921c65-0fb5-4d15-b303-c9b7d0021931-kube-api-access-jk95r\") pod \"cluster-samples-operator-665b6dd947-qsvvx\" (UID: \"1c921c65-0fb5-4d15-b303-c9b7d0021931\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.346684 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.347589 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvr6b\" (UniqueName: \"kubernetes.io/projected/74e4f428-25a3-45b9-959e-8a1eeb968cdf-kube-api-access-nvr6b\") pod \"apiserver-76f77b778f-9mrng\" (UID: \"74e4f428-25a3-45b9-959e-8a1eeb968cdf\") " pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.349471 4751 projected.go:194] Error preparing data for projected volume kube-api-access-pmptm for pod hostpath-provisioner/csi-hostpathplugin-69wfh: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.349554 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b1a71ab9-7855-40cf-9670-86381e49c4bc-kube-api-access-pmptm podName:b1a71ab9-7855-40cf-9670-86381e49c4bc nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.849531008 +0000 UTC m=+125.118863328 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-pmptm" (UniqueName: "kubernetes.io/projected/b1a71ab9-7855-40cf-9670-86381e49c4bc-kube-api-access-pmptm") pod "csi-hostpathplugin-69wfh" (UID: "b1a71ab9-7855-40cf-9670-86381e49c4bc") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.372553 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.372798 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.387881 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.390793 4751 projected.go:194] Error preparing data for projected volume kube-api-access-mcp7q for pod openshift-service-ca/service-ca-9c57cc56f-nx86z: failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.390895 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c0f92561-e30f-4e32-80d4-b463fd87de66-kube-api-access-mcp7q podName:c0f92561-e30f-4e32-80d4-b463fd87de66 nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.890873646 +0000 UTC m=+125.160205936 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mcp7q" (UniqueName: "kubernetes.io/projected/c0f92561-e30f-4e32-80d4-b463fd87de66-kube-api-access-mcp7q") pod "service-ca-9c57cc56f-nx86z" (UID: "c0f92561-e30f-4e32-80d4-b463fd87de66") : failed to sync configmap cache: timed out waiting for the condition Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.410277 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.416206 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.436579 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.438897 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.439082 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.939057299 +0000 UTC m=+125.208389589 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.439156 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.439827 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84j5b\" (UniqueName: \"kubernetes.io/projected/59c6efd1-70e9-45bc-8b87-94142678563d-kube-api-access-84j5b\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.439978 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.441361 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:16.941345037 +0000 UTC m=+125.210677327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.446980 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84j5b\" (UniqueName: \"kubernetes.io/projected/59c6efd1-70e9-45bc-8b87-94142678563d-kube-api-access-84j5b\") pod \"ingress-operator-5b745b69d9-k5qzx\" (UID: \"59c6efd1-70e9-45bc-8b87-94142678563d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.447390 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.454194 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rprdx\" (UniqueName: \"kubernetes.io/projected/c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05-kube-api-access-rprdx\") pod \"auto-csr-approver-29560328-lt56z\" (UID: \"c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05\") " pod="openshift-infra/auto-csr-approver-29560328-lt56z" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.467066 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.473710 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.477413 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6d776"] Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.487819 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.491014 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.513663 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.537863 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjxhz\" (UniqueName: \"kubernetes.io/projected/2b4bf065-6c73-43a0-b02f-a03188a64add-kube-api-access-xjxhz\") pod \"migrator-59844c95c7-g9fxd\" (UID: \"2b4bf065-6c73-43a0-b02f-a03188a64add\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.538568 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.538737 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.539295 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6qwkx"] Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.543592 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.543985 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.0439696 +0000 UTC m=+125.313301890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.545623 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.547201 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.553984 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.571255 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.574782 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.588466 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.597310 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6k5cn" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.635514 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.646847 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk"] Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.651734 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.652080 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560328-lt56z" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.653682 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.654161 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.154146428 +0000 UTC m=+125.423478718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.654234 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.654508 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.655310 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8lhg\" (UniqueName: \"kubernetes.io/projected/8545bd4e-a993-4f46-9d64-963a5f685a2d-kube-api-access-m8lhg\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.655467 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmdpq\" (UniqueName: \"kubernetes.io/projected/adb49d8b-968b-4fb1-8464-17853a08dda8-kube-api-access-mmdpq\") pod \"ingress-canary-nsnh2\" (UID: \"adb49d8b-968b-4fb1-8464-17853a08dda8\") " pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.656584 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh7cp\" (UniqueName: \"kubernetes.io/projected/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-kube-api-access-bh7cp\") pod \"package-server-manager-789f6589d5-xp482\" (UID: \"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.656839 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shm2f\" (UniqueName: \"kubernetes.io/projected/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-kube-api-access-shm2f\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.668346 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmdpq\" (UniqueName: \"kubernetes.io/projected/adb49d8b-968b-4fb1-8464-17853a08dda8-kube-api-access-mmdpq\") pod \"ingress-canary-nsnh2\" (UID: \"adb49d8b-968b-4fb1-8464-17853a08dda8\") " pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.669195 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh7cp\" (UniqueName: \"kubernetes.io/projected/82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e-kube-api-access-bh7cp\") pod \"package-server-manager-789f6589d5-xp482\" (UID: \"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.669993 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.673260 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shm2f\" (UniqueName: \"kubernetes.io/projected/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-kube-api-access-shm2f\") pod \"oauth-openshift-558db77b4-sfnlp\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.675790 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8lhg\" (UniqueName: \"kubernetes.io/projected/8545bd4e-a993-4f46-9d64-963a5f685a2d-kube-api-access-m8lhg\") pod \"packageserver-d55dfcdfc-rl4gt\" (UID: \"8545bd4e-a993-4f46-9d64-963a5f685a2d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.686400 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.690121 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.690319 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.716587 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd"] Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.733519 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.742708 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.757986 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.758196 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69gt4\" (UniqueName: \"kubernetes.io/projected/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-kube-api-access-69gt4\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.758229 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb49d\" (UniqueName: \"kubernetes.io/projected/c9118707-2746-44ed-bc4d-004312b53b5d-kube-api-access-pb49d\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.759296 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.259266825 +0000 UTC m=+125.528599115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.765364 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69gt4\" (UniqueName: \"kubernetes.io/projected/488dbfc7-cfc6-4c83-98d4-b2bc9f16969d-kube-api-access-69gt4\") pod \"service-ca-operator-777779d784-znlfm\" (UID: \"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.772762 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb49d\" (UniqueName: \"kubernetes.io/projected/c9118707-2746-44ed-bc4d-004312b53b5d-kube-api-access-pb49d\") pod \"kube-storage-version-migrator-operator-b67b599dd-c9jpp\" (UID: \"c9118707-2746-44ed-bc4d-004312b53b5d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.800355 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.812044 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx"] Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.812141 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.812435 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.818445 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.828699 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.861302 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.861391 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk6sg\" (UniqueName: \"kubernetes.io/projected/69dee6c1-1c2f-4d49-8955-ca4226b4306e-kube-api-access-fk6sg\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.861429 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmptm\" (UniqueName: \"kubernetes.io/projected/b1a71ab9-7855-40cf-9670-86381e49c4bc-kube-api-access-pmptm\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.861787 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.361773404 +0000 UTC m=+125.631105694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.867475 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.867694 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.879930 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmptm\" (UniqueName: \"kubernetes.io/projected/b1a71ab9-7855-40cf-9670-86381e49c4bc-kube-api-access-pmptm\") pod \"csi-hostpathplugin-69wfh\" (UID: \"b1a71ab9-7855-40cf-9670-86381e49c4bc\") " pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.880043 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.880991 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk6sg\" (UniqueName: \"kubernetes.io/projected/69dee6c1-1c2f-4d49-8955-ca4226b4306e-kube-api-access-fk6sg\") pod \"marketplace-operator-79b997595-sph8v\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.907001 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.916751 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.947025 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.952338 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nsnh2" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.960536 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm"] Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.965585 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.965754 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcp7q\" (UniqueName: \"kubernetes.io/projected/c0f92561-e30f-4e32-80d4-b463fd87de66-kube-api-access-mcp7q\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:16 crc kubenswrapper[4751]: E0316 00:08:16.971946 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.471925782 +0000 UTC m=+125.741258072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.987142 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcp7q\" (UniqueName: \"kubernetes.io/projected/c0f92561-e30f-4e32-80d4-b463fd87de66-kube-api-access-mcp7q\") pod \"service-ca-9c57cc56f-nx86z\" (UID: \"c0f92561-e30f-4e32-80d4-b463fd87de66\") " pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.987577 4751 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 16 00:08:16 crc kubenswrapper[4751]: I0316 00:08:16.996360 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-69wfh" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.005061 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6"] Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.067553 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.068048 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.568032319 +0000 UTC m=+125.837364599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.090752 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.090969 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.168569 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.169076 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.669060913 +0000 UTC m=+125.938393203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.178910 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.185564 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.191650 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f"] Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.210554 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zb5z5"] Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.224271 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-88ldb"] Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.270229 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.270741 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.770722956 +0000 UTC m=+126.040055256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.307053 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f"] Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.371575 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.371852 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.871828423 +0000 UTC m=+126.141160703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.372043 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.372522 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.872514363 +0000 UTC m=+126.141846653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.377647 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bd5b9"] Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.397808 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" event={"ID":"e55dc627-34bc-4853-ba79-146754ae58b5","Type":"ContainerStarted","Data":"8d87f9cb65af4bfb0a78f91272c5e006e7c17f937098f75ceb1b247688781efc"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.401544 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" event={"ID":"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa","Type":"ContainerStarted","Data":"8ebfaca053131272168e426555a0614a799cef4f96a2ec991db5745243ebcf78"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.401592 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" event={"ID":"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa","Type":"ContainerStarted","Data":"ad62594a3a080574963a600a65cd1972f73540e32010764471048e0a226fc6f5"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.403173 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" event={"ID":"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256","Type":"ContainerStarted","Data":"2cbc3e4c1ede178a2ee896c0fab720d4de2c789d958538fdc94c96e7dba5589f"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.405316 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6d776" event={"ID":"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75","Type":"ContainerStarted","Data":"787edb3323950be6376e1944dd5f84315023d617902f2426f90fe0126a857387"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.405342 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6d776" event={"ID":"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75","Type":"ContainerStarted","Data":"9988946d4abd3cb885975bba7cfcaffc42805787a0eca78c54187e0c9c9b11c8"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.406455 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" event={"ID":"3b5183fc-18b0-4aa5-9d1c-de26657da9f7","Type":"ContainerStarted","Data":"184d007961de292d9cf9d8a823bd6770ce77eff3a749f3c259e2034121cd1b21"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.407362 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" event={"ID":"a932e9e3-ca99-4af3-a9e6-8888e31be60d","Type":"ContainerStarted","Data":"e1a9cfb82bacfb6768193b57fd72e38f5a0d0ff38dbb7d353693a6255289305b"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.408254 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" event={"ID":"5620fac3-59f5-49e4-9125-501353fbf836","Type":"ContainerStarted","Data":"c96ab91f6120a0f94f709b30b584e86e7b81311ed636dc7c5a1bcf9001fc67a0"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.409085 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" event={"ID":"c1ee3303-f745-4058-8927-046b07740c33","Type":"ContainerStarted","Data":"30d84c7a9999bddacac9458b694992d0f4719a7fa9bd089499464b665da69f76"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.411370 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" event={"ID":"b6d655c6-0506-41b7-b032-842ff9630e41","Type":"ContainerStarted","Data":"fd83c12283a57c413ed17ed59f74a01b80eafaa4ab90432caf271617bb235fd5"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.417354 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jpvk6" event={"ID":"a792c668-ee31-4bd8-a122-37bf8b48f37f","Type":"ContainerStarted","Data":"fcaff13f4ac0f4c5faaa78ad4ca43701c45e6ba924e8d9df1d62628824288eaf"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.417410 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jpvk6" event={"ID":"a792c668-ee31-4bd8-a122-37bf8b48f37f","Type":"ContainerStarted","Data":"16d394961aff6a1a3f4729f6d6dc920a612b67c7828f014984ba717deeb96b51"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.423858 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" event={"ID":"270222df-3862-47e2-a363-c69546415e18","Type":"ContainerStarted","Data":"fdf4c04356ae19f9fcd8a413e67c6f7e3b34ca3f332dc69de323101ddbbba44a"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.431287 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" event={"ID":"b60da8df-4808-44ec-8e43-bc5e7b56079d","Type":"ContainerStarted","Data":"f5122b22768eef86e1fc429bcb01d1ea1fb0392d8e9bdd536bc75b73a49a5153"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.443382 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" event={"ID":"1e707425-3377-458e-a331-6af00d1e989c","Type":"ContainerStarted","Data":"1abb514521ad62fff421958261bb983792e5dc4e12e49b162c5567d8cf38276e"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.443427 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" event={"ID":"1e707425-3377-458e-a331-6af00d1e989c","Type":"ContainerStarted","Data":"f3afb0bd9d401df269796ada04c74396cd1bcb9ebbd5c4426a728e003f6da8c3"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.443437 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" event={"ID":"1e707425-3377-458e-a331-6af00d1e989c","Type":"ContainerStarted","Data":"dcd4e75c262b306018b88b8f2abd02956675d539bcb27538ce25cbdbae7e907e"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.445239 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" event={"ID":"2d148ff3-a331-4c35-925c-3e3d0124a702","Type":"ContainerStarted","Data":"3ae553415562a1b3af5b9757369d4d6de26e78462e7190cb5e13e7f6a0c8b00f"} Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.454780 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-njjzb" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.473711 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.474000 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:17.973985581 +0000 UTC m=+126.243317871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.493798 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.576385 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.578403 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.078389737 +0000 UTC m=+126.347722027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.618586 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:17 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:17 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:17 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.618643 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:17 crc kubenswrapper[4751]: W0316 00:08:17.640537 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfe3c553_b411_4074_b83a_c8eba2518089.slice/crio-cddc879c34cdd0465d0725c9d0cae79264aed2f0b02bb73b4a26b463f9ab1e0e WatchSource:0}: Error finding container cddc879c34cdd0465d0725c9d0cae79264aed2f0b02bb73b4a26b463f9ab1e0e: Status 404 returned error can't find the container with id cddc879c34cdd0465d0725c9d0cae79264aed2f0b02bb73b4a26b463f9ab1e0e Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.680646 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.680929 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.180904526 +0000 UTC m=+126.450236816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.681206 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.681587 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.181573726 +0000 UTC m=+126.450906016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.738835 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-jszhl" podStartSLOduration=51.73881489 podStartE2EDuration="51.73881489s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:17.738036076 +0000 UTC m=+126.007368366" watchObservedRunningTime="2026-03-16 00:08:17.73881489 +0000 UTC m=+126.008147170" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.739049 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" podStartSLOduration=50.739042717 podStartE2EDuration="50.739042717s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:17.710458941 +0000 UTC m=+125.979791231" watchObservedRunningTime="2026-03-16 00:08:17.739042717 +0000 UTC m=+126.008374997" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.813166 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-njjzb" podStartSLOduration=51.813143165 podStartE2EDuration="51.813143165s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:17.805408453 +0000 UTC m=+126.074740743" watchObservedRunningTime="2026-03-16 00:08:17.813143165 +0000 UTC m=+126.082475455" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.823420 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.824127 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.324080562 +0000 UTC m=+126.593412852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.912858 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29560320-xlv4n" podStartSLOduration=51.91284439 podStartE2EDuration="51.91284439s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:17.911057386 +0000 UTC m=+126.180389676" watchObservedRunningTime="2026-03-16 00:08:17.91284439 +0000 UTC m=+126.182176680" Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.925977 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:17 crc kubenswrapper[4751]: E0316 00:08:17.926560 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.42654647 +0000 UTC m=+126.695878760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.963900 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx"] Mar 16 00:08:17 crc kubenswrapper[4751]: I0316 00:08:17.965304 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-x9l6q" podStartSLOduration=51.96528296 podStartE2EDuration="51.96528296s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:17.948415375 +0000 UTC m=+126.217747665" watchObservedRunningTime="2026-03-16 00:08:17.96528296 +0000 UTC m=+126.234615250" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.027258 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.027502 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.527472942 +0000 UTC m=+126.796805232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.027729 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.028065 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.528053609 +0000 UTC m=+126.797385899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.045354 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" podStartSLOduration=8.045337776 podStartE2EDuration="8.045337776s" podCreationTimestamp="2026-03-16 00:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:18.043007707 +0000 UTC m=+126.312339997" watchObservedRunningTime="2026-03-16 00:08:18.045337776 +0000 UTC m=+126.314670066" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.083246 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.087057 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w"] Mar 16 00:08:18 crc kubenswrapper[4751]: W0316 00:08:18.121722 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59c6efd1_70e9_45bc_8b87_94142678563d.slice/crio-759d96e4a8506404239dda1136f5a2b40df1bb74a6cf62aa5901b7e77b8d2430 WatchSource:0}: Error finding container 759d96e4a8506404239dda1136f5a2b40df1bb74a6cf62aa5901b7e77b8d2430: Status 404 returned error can't find the container with id 759d96e4a8506404239dda1136f5a2b40df1bb74a6cf62aa5901b7e77b8d2430 Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.132280 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.132587 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.632572988 +0000 UTC m=+126.901905268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: W0316 00:08:18.140059 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2676895_4d0e_4099_9d49_abd723369982.slice/crio-18da8a0a2bd065aca2ba6dd35af1ea1cc81e950e018d28cd69617a536fcf7ce1 WatchSource:0}: Error finding container 18da8a0a2bd065aca2ba6dd35af1ea1cc81e950e018d28cd69617a536fcf7ce1: Status 404 returned error can't find the container with id 18da8a0a2bd065aca2ba6dd35af1ea1cc81e950e018d28cd69617a536fcf7ce1 Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.151203 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6k5cn"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.170969 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mrm4x"] Mar 16 00:08:18 crc kubenswrapper[4751]: W0316 00:08:18.212754 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2394c21b_2d69_4698_8e58_d0198ec77558.slice/crio-9852cec68ed22e024ce3494436bc9eb4ad4a4d4b129178adef187240c99ba5aa WatchSource:0}: Error finding container 9852cec68ed22e024ce3494436bc9eb4ad4a4d4b129178adef187240c99ba5aa: Status 404 returned error can't find the container with id 9852cec68ed22e024ce3494436bc9eb4ad4a4d4b129178adef187240c99ba5aa Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.233287 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.233694 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.733681535 +0000 UTC m=+127.003013825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.266848 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.312267 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dklns" podStartSLOduration=52.312247997 podStartE2EDuration="52.312247997s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:18.304298749 +0000 UTC m=+126.573631049" watchObservedRunningTime="2026-03-16 00:08:18.312247997 +0000 UTC m=+126.581580287" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.312770 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt"] Mar 16 00:08:18 crc kubenswrapper[4751]: W0316 00:08:18.322701 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8545bd4e_a993_4f46_9d64_963a5f685a2d.slice/crio-4a441d2af133f40252e09efccecf641a773ccf6c67bb1a65024768a01cfd5fde WatchSource:0}: Error finding container 4a441d2af133f40252e09efccecf641a773ccf6c67bb1a65024768a01cfd5fde: Status 404 returned error can't find the container with id 4a441d2af133f40252e09efccecf641a773ccf6c67bb1a65024768a01cfd5fde Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.333949 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.334486 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.834464132 +0000 UTC m=+127.103796412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.342057 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560328-lt56z"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.347894 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.352059 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.373511 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.390929 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9mrng"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.392930 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.421793 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-p56l4" podStartSLOduration=8.421777836 podStartE2EDuration="8.421777836s" podCreationTimestamp="2026-03-16 00:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:18.420741145 +0000 UTC m=+126.690073435" watchObservedRunningTime="2026-03-16 00:08:18.421777836 +0000 UTC m=+126.691110126" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.441762 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.444212 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:18.944194128 +0000 UTC m=+127.213526418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.510023 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:18 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:18 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:18 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.510694 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.533060 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560328-lt56z" event={"ID":"c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05","Type":"ContainerStarted","Data":"4898c76f3dc20881acc5936dd31aa7be9d7710f3784cc9d50d190e8e93058c1d"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.533237 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sfnlp"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.533328 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-znlfm"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.533395 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" event={"ID":"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e","Type":"ContainerStarted","Data":"1e7e9f27b65999f7e92ddabaeb822f11251680328dfa3a137eb9174664752496"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.533461 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" event={"ID":"122e1c85-74cb-41e2-bc9f-aa5c3ac4a5aa","Type":"ContainerStarted","Data":"f623bbb8d804f8157438ac8b04acd73849241ae81bad7cdb6b5109d76603c2f7"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.534704 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sph8v"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.534780 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-69wfh"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.539078 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" event={"ID":"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1","Type":"ContainerStarted","Data":"641c4cae1b8a82343034d34a2fbabd2c937280bd28e811464a89fe6eaee99a0f"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.539255 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" event={"ID":"a3f6093c-dd51-45d1-9d18-0a80bac4fdb1","Type":"ContainerStarted","Data":"0fd83dac91defbaef8bbefc9500850787066da6604cf934132321708bc12bc1e"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.540155 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.553761 4751 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-rfk2f container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.553815 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" podUID="a3f6093c-dd51-45d1-9d18-0a80bac4fdb1" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.553946 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.554195 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.05416915 +0000 UTC m=+127.323501440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.554441 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.555325 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.055303994 +0000 UTC m=+127.324636284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.555376 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nx86z"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.573049 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nsnh2"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.574552 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" event={"ID":"a8fdc932-6670-45f5-9f69-c15eaae40bc4","Type":"ContainerStarted","Data":"1cced3cef8665135e290460c0b3834e20fa11d9913decb63fdeb4cb6588d883b"} Mar 16 00:08:18 crc kubenswrapper[4751]: W0316 00:08:18.585935 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1a71ab9_7855_40cf_9670_86381e49c4bc.slice/crio-15d52b0e6494d2628402d44911483fba95a2fa41a321949ae9b258a4f74d8512 WatchSource:0}: Error finding container 15d52b0e6494d2628402d44911483fba95a2fa41a321949ae9b258a4f74d8512: Status 404 returned error can't find the container with id 15d52b0e6494d2628402d44911483fba95a2fa41a321949ae9b258a4f74d8512 Mar 16 00:08:18 crc kubenswrapper[4751]: W0316 00:08:18.589877 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0f92561_e30f_4e32_80d4_b463fd87de66.slice/crio-7c1b9fba5b8a8448636bd66c267df694d5e024cea745a912e2994a5330bcf15c WatchSource:0}: Error finding container 7c1b9fba5b8a8448636bd66c267df694d5e024cea745a912e2994a5330bcf15c: Status 404 returned error can't find the container with id 7c1b9fba5b8a8448636bd66c267df694d5e024cea745a912e2994a5330bcf15c Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.590065 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" event={"ID":"59c6efd1-70e9-45bc-8b87-94142678563d","Type":"ContainerStarted","Data":"55eb8685df316cd096fc3ac3ca37fa0ada91d0ea68761e9d9f7f66c98d3ec54b"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.590133 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" event={"ID":"59c6efd1-70e9-45bc-8b87-94142678563d","Type":"ContainerStarted","Data":"759d96e4a8506404239dda1136f5a2b40df1bb74a6cf62aa5901b7e77b8d2430"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.608341 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd" event={"ID":"2b4bf065-6c73-43a0-b02f-a03188a64add","Type":"ContainerStarted","Data":"9af95b0de8300f434ee652049fd9752b0364d5f85051e4276506c206dcb29625"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.628946 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" event={"ID":"2d148ff3-a331-4c35-925c-3e3d0124a702","Type":"ContainerStarted","Data":"18492b634c46f7aa8b85e983b814c8acce4ad8af00c8c52b1b68193377a80281"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.631286 4751 generic.go:334] "Generic (PLEG): container finished" podID="3b5183fc-18b0-4aa5-9d1c-de26657da9f7" containerID="166fe0f3c532f49fe8083bb620dc376ac25f9c8f1a1984ba172a9f372b2b4be4" exitCode=0 Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.632344 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" event={"ID":"3b5183fc-18b0-4aa5-9d1c-de26657da9f7","Type":"ContainerDied","Data":"166fe0f3c532f49fe8083bb620dc376ac25f9c8f1a1984ba172a9f372b2b4be4"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.635486 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" event={"ID":"270222df-3862-47e2-a363-c69546415e18","Type":"ContainerStarted","Data":"36f6b8df1ebf6d8b145e1d2740b4bfa72be0bd6699c8d9d844166d30df3cfd5b"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.635518 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" event={"ID":"270222df-3862-47e2-a363-c69546415e18","Type":"ContainerStarted","Data":"b93322dad23c389731e2eb9d3c9aa3ece3c8edf03e4189f218b8965e0f9ccfd4"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.644271 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" event={"ID":"74e4f428-25a3-45b9-959e-8a1eeb968cdf","Type":"ContainerStarted","Data":"053ef6931a13a6bd7deb40da74ba37ad55f7c26b9272f577d455ad0ed2526d4a"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.656000 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.656157 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.156133293 +0000 UTC m=+127.425465583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.656320 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.657861 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.157842654 +0000 UTC m=+127.427174934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.662948 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g5frw" podStartSLOduration=52.662925116 podStartE2EDuration="52.662925116s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:18.621244008 +0000 UTC m=+126.890576298" watchObservedRunningTime="2026-03-16 00:08:18.662925116 +0000 UTC m=+126.932257406" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.671471 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" event={"ID":"a932e9e3-ca99-4af3-a9e6-8888e31be60d","Type":"ContainerStarted","Data":"9a5551f66812ea5c6dd90d294a14fbcb7b694426454e3f4ac7ac80ceaaf7a653"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.671515 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" event={"ID":"a932e9e3-ca99-4af3-a9e6-8888e31be60d","Type":"ContainerStarted","Data":"52a364f61941cb7797682ad0734d5a18f47d7a633255eac50eddba58dc5f0bdc"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.687849 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" event={"ID":"b60da8df-4808-44ec-8e43-bc5e7b56079d","Type":"ContainerStarted","Data":"65c0ac7105248640c6bc1440ed88bad41a84ae3d406b22cae971006248b43635"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.701029 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vg885" podStartSLOduration=52.701008306 podStartE2EDuration="52.701008306s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:18.664913745 +0000 UTC m=+126.934246045" watchObservedRunningTime="2026-03-16 00:08:18.701008306 +0000 UTC m=+126.970340596" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.715314 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" event={"ID":"8545bd4e-a993-4f46-9d64-963a5f685a2d","Type":"ContainerStarted","Data":"4a441d2af133f40252e09efccecf641a773ccf6c67bb1a65024768a01cfd5fde"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.716220 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.719789 4751 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rl4gt container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" start-of-body= Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.719824 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" podUID="8545bd4e-a993-4f46-9d64-963a5f685a2d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.722651 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" event={"ID":"c9118707-2746-44ed-bc4d-004312b53b5d","Type":"ContainerStarted","Data":"d4d9c26c1378fa95532bbc9392632e162dcc013b9d2987ff28eb1c911e2f415f"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.722684 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" event={"ID":"c9118707-2746-44ed-bc4d-004312b53b5d","Type":"ContainerStarted","Data":"aaee91b197e4ca24ddce5717c409538e2bd97e8a30712d6cd6ce37109abcc46e"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.725333 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bd5b9" event={"ID":"cfe3c553-b411-4074-b83a-c8eba2518089","Type":"ContainerStarted","Data":"af9b44f979ecdb34f850eb3f843f422b6e47398a43ed58da091468e1c21cc5c9"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.725371 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bd5b9" event={"ID":"cfe3c553-b411-4074-b83a-c8eba2518089","Type":"ContainerStarted","Data":"cddc879c34cdd0465d0725c9d0cae79264aed2f0b02bb73b4a26b463f9ab1e0e"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.729512 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6k5cn" event={"ID":"2394c21b-2d69-4698-8e58-d0198ec77558","Type":"ContainerStarted","Data":"9852cec68ed22e024ce3494436bc9eb4ad4a4d4b129178adef187240c99ba5aa"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.730240 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6k5cn" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.732885 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-6k5cn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.732925 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6k5cn" podUID="2394c21b-2d69-4698-8e58-d0198ec77558" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.738357 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" event={"ID":"a230e2c9-cb14-4f26-8552-d81e42a7fcc2","Type":"ContainerStarted","Data":"e47b68f52727f89089aa66a0fdef1a05d2a84565304320f6c15017990295e584"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.743859 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" event={"ID":"5620fac3-59f5-49e4-9125-501353fbf836","Type":"ContainerStarted","Data":"4ff53e00dd4533a2ed491e7fe81afa1abd2fb9f31b56659af18aa0764f54f5bc"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.757189 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.757490 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" event={"ID":"c1ee3303-f745-4058-8927-046b07740c33","Type":"ContainerStarted","Data":"6e658e57de21d53ddcd5a4aa0880e33a6cb3b5355c7697b2b07e183476ec3f1f"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.757524 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" event={"ID":"c1ee3303-f745-4058-8927-046b07740c33","Type":"ContainerStarted","Data":"1418833c9021b78fc4f9f69f16f76fde4e5216946c207dafb29ba24024e65736"} Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.758461 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.258441845 +0000 UTC m=+127.527774135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.763125 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" event={"ID":"c41e2fa1-c06c-4bdb-8a7b-11d94d47b256","Type":"ContainerStarted","Data":"7f2d4716c108737bb7e786e41884c3f323014000ba8257cd6505c4ad30bc6f73"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.765124 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" event={"ID":"c2676895-4d0e-4099-9d49-abd723369982","Type":"ContainerStarted","Data":"18da8a0a2bd065aca2ba6dd35af1ea1cc81e950e018d28cd69617a536fcf7ce1"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.779151 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-t2bcd" podStartSLOduration=52.779132725 podStartE2EDuration="52.779132725s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:18.778423874 +0000 UTC m=+127.047756154" watchObservedRunningTime="2026-03-16 00:08:18.779132725 +0000 UTC m=+127.048465015" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.782194 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" event={"ID":"e55dc627-34bc-4853-ba79-146754ae58b5","Type":"ContainerStarted","Data":"47ecea2ad10c8ec1bc5c4872ba532ac102c36af7ad3a61834c14bfcd6dbd4f40"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.783066 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.786668 4751 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fwd6f container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.786710 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" podUID="e55dc627-34bc-4853-ba79-146754ae58b5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.790044 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6d776" event={"ID":"bbbc093a-f5f6-4db8-89ae-1c3224e6bf75","Type":"ContainerStarted","Data":"71432d3b6c2f8384a5ec3a6be9cb6c31426e524899b9245a62e255e59e0c02a9"} Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.790078 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6d776" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.810756 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-hz7xk"] Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.823326 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vl6bg" podStartSLOduration=52.823305007 podStartE2EDuration="52.823305007s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:18.82173572 +0000 UTC m=+127.091068000" watchObservedRunningTime="2026-03-16 00:08:18.823305007 +0000 UTC m=+127.092637297" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.866287 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.885847 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.385829599 +0000 UTC m=+127.655161889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.887183 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" podStartSLOduration=51.887168279 podStartE2EDuration="51.887168279s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:18.86547525 +0000 UTC m=+127.134807540" watchObservedRunningTime="2026-03-16 00:08:18.887168279 +0000 UTC m=+127.156500569" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.949004 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-jpvk6" podStartSLOduration=52.94898041 podStartE2EDuration="52.94898041s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:18.928698423 +0000 UTC m=+127.198030713" watchObservedRunningTime="2026-03-16 00:08:18.94898041 +0000 UTC m=+127.218312700" Mar 16 00:08:18 crc kubenswrapper[4751]: I0316 00:08:18.968786 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:18 crc kubenswrapper[4751]: E0316 00:08:18.969135 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.469120073 +0000 UTC m=+127.738452363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.031446 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.032152 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.041957 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.060859 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" podStartSLOduration=53.060839819 podStartE2EDuration="53.060839819s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.060658453 +0000 UTC m=+127.329990743" watchObservedRunningTime="2026-03-16 00:08:19.060839819 +0000 UTC m=+127.330172099" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.069981 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.071144 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.571124967 +0000 UTC m=+127.840457257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.100552 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" podStartSLOduration=52.100534917 podStartE2EDuration="52.100534917s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.098188457 +0000 UTC m=+127.367520737" watchObservedRunningTime="2026-03-16 00:08:19.100534917 +0000 UTC m=+127.369867207" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.139813 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr4v6" podStartSLOduration=53.139796393 podStartE2EDuration="53.139796393s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.138519414 +0000 UTC m=+127.407851704" watchObservedRunningTime="2026-03-16 00:08:19.139796393 +0000 UTC m=+127.409128683" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.171082 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.171258 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.671233354 +0000 UTC m=+127.940565644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.171422 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.171772 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.671757799 +0000 UTC m=+127.941090089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.182423 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6d776" podStartSLOduration=9.182405058 podStartE2EDuration="9.182405058s" podCreationTimestamp="2026-03-16 00:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.179033977 +0000 UTC m=+127.448366267" watchObservedRunningTime="2026-03-16 00:08:19.182405058 +0000 UTC m=+127.451737348" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.222868 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c9jpp" podStartSLOduration=53.222852249 podStartE2EDuration="53.222852249s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.221923611 +0000 UTC m=+127.491255901" watchObservedRunningTime="2026-03-16 00:08:19.222852249 +0000 UTC m=+127.492184539" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.260318 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-6qwkx" podStartSLOduration=53.26029667 podStartE2EDuration="53.26029667s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.259654631 +0000 UTC m=+127.528986921" watchObservedRunningTime="2026-03-16 00:08:19.26029667 +0000 UTC m=+127.529628960" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.272721 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.273081 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.773036751 +0000 UTC m=+128.042369041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.340731 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-bd5b9" podStartSLOduration=53.340709357 podStartE2EDuration="53.340709357s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.339527822 +0000 UTC m=+127.608860102" watchObservedRunningTime="2026-03-16 00:08:19.340709357 +0000 UTC m=+127.610041647" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.374068 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.374404 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.874391496 +0000 UTC m=+128.143723786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.386708 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nh2cx" podStartSLOduration=53.386675104 podStartE2EDuration="53.386675104s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.386306853 +0000 UTC m=+127.655639143" watchObservedRunningTime="2026-03-16 00:08:19.386675104 +0000 UTC m=+127.656007394" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.474744 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.474950 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.974919565 +0000 UTC m=+128.244251865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.475324 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.475732 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:19.97572211 +0000 UTC m=+128.245054400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.496978 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:19 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:19 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:19 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.497058 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.499880 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wdvsd" podStartSLOduration=53.499866342 podStartE2EDuration="53.499866342s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.499226983 +0000 UTC m=+127.768559283" watchObservedRunningTime="2026-03-16 00:08:19.499866342 +0000 UTC m=+127.769198632" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.576975 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.577210 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.077179857 +0000 UTC m=+128.346512147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.577274 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.577852 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.077837967 +0000 UTC m=+128.347170247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.672708 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34180: no serving certificate available for the kubelet" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.680040 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.681203 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.18116742 +0000 UTC m=+128.450499710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.683515 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.684349 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.184336245 +0000 UTC m=+128.453668535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.746380 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34184: no serving certificate available for the kubelet" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.749274 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w4zmm" podStartSLOduration=52.749259429 podStartE2EDuration="52.749259429s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.747540687 +0000 UTC m=+128.016872977" watchObservedRunningTime="2026-03-16 00:08:19.749259429 +0000 UTC m=+128.018591709" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.780527 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" podStartSLOduration=53.780510075 podStartE2EDuration="53.780510075s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.78035678 +0000 UTC m=+128.049689070" watchObservedRunningTime="2026-03-16 00:08:19.780510075 +0000 UTC m=+128.049842365" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.784245 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.784666 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.284649478 +0000 UTC m=+128.553981778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.805397 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-69wfh" event={"ID":"b1a71ab9-7855-40cf-9670-86381e49c4bc","Type":"ContainerStarted","Data":"15d52b0e6494d2628402d44911483fba95a2fa41a321949ae9b258a4f74d8512"} Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.811753 4751 generic.go:334] "Generic (PLEG): container finished" podID="74e4f428-25a3-45b9-959e-8a1eeb968cdf" containerID="c7386a90ed237a3273c9174535407442e66848b3a4d50bf5166c70a59437ef75" exitCode=0 Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.811820 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" event={"ID":"74e4f428-25a3-45b9-959e-8a1eeb968cdf","Type":"ContainerDied","Data":"c7386a90ed237a3273c9174535407442e66848b3a4d50bf5166c70a59437ef75"} Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.814834 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" event={"ID":"c2676895-4d0e-4099-9d49-abd723369982","Type":"ContainerStarted","Data":"7ce477121853d95f45d375e8cd1901a4a918606253974a7d9b47ee82aef01a7e"} Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.845422 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" event={"ID":"a8fdc932-6670-45f5-9f69-c15eaae40bc4","Type":"ContainerStarted","Data":"dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b"} Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.846765 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.850418 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-zb5z5" podStartSLOduration=52.850408637 podStartE2EDuration="52.850408637s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.848917123 +0000 UTC m=+128.118249413" watchObservedRunningTime="2026-03-16 00:08:19.850408637 +0000 UTC m=+128.119740927" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.861828 4751 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mrm4x container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.861893 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" podUID="a8fdc932-6670-45f5-9f69-c15eaae40bc4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.863092 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34190: no serving certificate available for the kubelet" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.883234 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" podStartSLOduration=53.883087165 podStartE2EDuration="53.883087165s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.880929841 +0000 UTC m=+128.150262131" watchObservedRunningTime="2026-03-16 00:08:19.883087165 +0000 UTC m=+128.152419445" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.885464 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.888121 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d528n" event={"ID":"a230e2c9-cb14-4f26-8552-d81e42a7fcc2","Type":"ContainerStarted","Data":"fb39be88e9ba9940ec903cd86c913a794f40ff2f33b224e3e891d933a9750cec"} Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.888294 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.388279521 +0000 UTC m=+128.657611811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.945919 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-74gwk" podStartSLOduration=53.945900186 podStartE2EDuration="53.945900186s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.910916339 +0000 UTC m=+128.180248629" watchObservedRunningTime="2026-03-16 00:08:19.945900186 +0000 UTC m=+128.215232476" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.955538 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34206: no serving certificate available for the kubelet" Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.966323 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd" event={"ID":"2b4bf065-6c73-43a0-b02f-a03188a64add","Type":"ContainerStarted","Data":"1240cbd87e3fa4b26c77ea8c9ca06c4fa36a8e8aa7a853dfe47d3d7c09b73b17"} Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.967308 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd" event={"ID":"2b4bf065-6c73-43a0-b02f-a03188a64add","Type":"ContainerStarted","Data":"6d2ebd68f2b80bec0e58698bc4f3e115e4d64305530ae6b499afa70432cc9795"} Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.977270 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" event={"ID":"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d","Type":"ContainerStarted","Data":"caeba174e43aaaac261e8f59aa4ed24c677f181e263b040afdfda2acca9fe585"} Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.977316 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" event={"ID":"488dbfc7-cfc6-4c83-98d4-b2bc9f16969d","Type":"ContainerStarted","Data":"a007808dd5b877b26e4629a3c521978f6f100548dcf38178ac7708d19531191a"} Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.986728 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.986825 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6k5cn" podStartSLOduration=53.986809271 podStartE2EDuration="53.986809271s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:19.986556063 +0000 UTC m=+128.255888353" watchObservedRunningTime="2026-03-16 00:08:19.986809271 +0000 UTC m=+128.256141561" Mar 16 00:08:19 crc kubenswrapper[4751]: E0316 00:08:19.987674 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.487656456 +0000 UTC m=+128.756988746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:19 crc kubenswrapper[4751]: I0316 00:08:19.996891 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" event={"ID":"8545bd4e-a993-4f46-9d64-963a5f685a2d","Type":"ContainerStarted","Data":"f4c660d0bc94abfb80f7b38499e9e65c388d4a8040180cd081c98a344df53d74"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.024734 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" podStartSLOduration=53.024717546 podStartE2EDuration="53.024717546s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.02051428 +0000 UTC m=+128.289846570" watchObservedRunningTime="2026-03-16 00:08:20.024717546 +0000 UTC m=+128.294049836" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.026912 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" event={"ID":"3b5183fc-18b0-4aa5-9d1c-de26657da9f7","Type":"ContainerStarted","Data":"914bec563d6eaca900d9527a87587d7a5ec8927c2a235a8b99fd6b794dce04f2"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.026991 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.057341 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34220: no serving certificate available for the kubelet" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.064406 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" event={"ID":"69dee6c1-1c2f-4d49-8955-ca4226b4306e","Type":"ContainerStarted","Data":"bc4a634b7a61211753a19f21449ebaa9564426678f0caac161542bf34aafd110"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.064456 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" event={"ID":"69dee6c1-1c2f-4d49-8955-ca4226b4306e","Type":"ContainerStarted","Data":"f38fceacb6bd0d734314c04b36462819587a88449c72583b762c2fe4773dd5a4"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.065299 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.066738 4751 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sph8v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.066771 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.076697 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nsnh2" event={"ID":"adb49d8b-968b-4fb1-8464-17853a08dda8","Type":"ContainerStarted","Data":"b22101dd3887f970fac0680cc5ba5b922d27adef3cc5b3ec459f1810bd2264b9"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.076738 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nsnh2" event={"ID":"adb49d8b-968b-4fb1-8464-17853a08dda8","Type":"ContainerStarted","Data":"daf22d474e35e5fc3a006b2ffa283eb90e718e22e465c8476f0cd51767180127"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.092666 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" event={"ID":"59c6efd1-70e9-45bc-8b87-94142678563d","Type":"ContainerStarted","Data":"dd6ed5ac9db3e3556c498d587c5849cb0d76b8e42817bb0dc31fcd584b7dc32a"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.095024 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.097558 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.597544936 +0000 UTC m=+128.866877226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.110038 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jlhhf" podStartSLOduration=54.110021739 podStartE2EDuration="54.110021739s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.108505624 +0000 UTC m=+128.377837914" watchObservedRunningTime="2026-03-16 00:08:20.110021739 +0000 UTC m=+128.379354029" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.111155 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" event={"ID":"1c921c65-0fb5-4d15-b303-c9b7d0021931","Type":"ContainerStarted","Data":"a1a8d8f7c5aba29cffb7e7eb320d18f331ccef689b893f24ee859f2ce102145e"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.111216 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" event={"ID":"1c921c65-0fb5-4d15-b303-c9b7d0021931","Type":"ContainerStarted","Data":"7202e50e38f417cf3c8f301b2e0c21ce7320b78767a24ef9afa6ac964c784fbb"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.111231 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" event={"ID":"1c921c65-0fb5-4d15-b303-c9b7d0021931","Type":"ContainerStarted","Data":"c76126a4bcd359110b4f17886b50eff5420a455c3086d80c53e7487761037cc2"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.113794 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" event={"ID":"c0f92561-e30f-4e32-80d4-b463fd87de66","Type":"ContainerStarted","Data":"1547b56afd60feb3731f5e3d1b5d13d82906d0711a1afad0214d1f8ccd3acf81"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.113821 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" event={"ID":"c0f92561-e30f-4e32-80d4-b463fd87de66","Type":"ContainerStarted","Data":"7c1b9fba5b8a8448636bd66c267df694d5e024cea745a912e2994a5330bcf15c"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.122704 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6k5cn" event={"ID":"2394c21b-2d69-4698-8e58-d0198ec77558","Type":"ContainerStarted","Data":"ead04c020db71627e03dd298b1f4eb627b00a884b80f3cabbc54f13399b9282f"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.125183 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-6k5cn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.125234 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6k5cn" podUID="2394c21b-2d69-4698-8e58-d0198ec77558" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.133263 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" event={"ID":"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e","Type":"ContainerStarted","Data":"48950f0166254b86eb29f4811e21aa20427e36085b2c4e6e395d9782c9ab2511"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.133317 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" event={"ID":"82b3d3c7-3d6c-4d40-bb41-5a5ae972e64e","Type":"ContainerStarted","Data":"74f7cc0204df5f14435bac2cbce08766eb6d7560fc197959b814b62418d6b0be"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.133459 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.145794 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9fxd" podStartSLOduration=54.14577664 podStartE2EDuration="54.14577664s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.144953835 +0000 UTC m=+128.414286125" watchObservedRunningTime="2026-03-16 00:08:20.14577664 +0000 UTC m=+128.415108930" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.161400 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" event={"ID":"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec","Type":"ContainerStarted","Data":"741b041b7d841e9e364dbef29a97a96bc40b6cdb95edc0f376a8cf1cc664ad94"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.161447 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" event={"ID":"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec","Type":"ContainerStarted","Data":"d7ff99d1f527cf26cc2ce1c8c78a12f0a684da43c048147f7b54109ddd0bc364"} Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.162229 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" podUID="9c56ee92-a6f5-46ec-8076-157268b354c7" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" gracePeriod=30 Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.177262 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fwd6f" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.179388 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rfk2f" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.187868 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34236: no serving certificate available for the kubelet" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.188760 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tj4rk" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.196667 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.196938 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.69689482 +0000 UTC m=+128.966227110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.201607 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.209643 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.709610501 +0000 UTC m=+128.978942791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.233068 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" podStartSLOduration=54.233048633 podStartE2EDuration="54.233048633s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.230180497 +0000 UTC m=+128.499512787" watchObservedRunningTime="2026-03-16 00:08:20.233048633 +0000 UTC m=+128.502380923" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.262411 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5qzx" podStartSLOduration=54.262393431 podStartE2EDuration="54.262393431s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.261587617 +0000 UTC m=+128.530919907" watchObservedRunningTime="2026-03-16 00:08:20.262393431 +0000 UTC m=+128.531725721" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.306614 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.308141 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.80812317 +0000 UTC m=+129.077455460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.353371 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" podStartSLOduration=53.353346114 podStartE2EDuration="53.353346114s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.310473911 +0000 UTC m=+128.579806201" watchObservedRunningTime="2026-03-16 00:08:20.353346114 +0000 UTC m=+128.622678414" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.398192 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" podStartSLOduration=54.398173446 podStartE2EDuration="54.398173446s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.394766764 +0000 UTC m=+128.664099054" watchObservedRunningTime="2026-03-16 00:08:20.398173446 +0000 UTC m=+128.667505736" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.408864 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.409325 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:20.909305269 +0000 UTC m=+129.178637629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.423041 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-znlfm" podStartSLOduration=53.42302119 podStartE2EDuration="53.42302119s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.419806914 +0000 UTC m=+128.689139204" watchObservedRunningTime="2026-03-16 00:08:20.42302119 +0000 UTC m=+128.692353470" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.447606 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34250: no serving certificate available for the kubelet" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.474735 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-nsnh2" podStartSLOduration=10.474712558 podStartE2EDuration="10.474712558s" podCreationTimestamp="2026-03-16 00:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.472925214 +0000 UTC m=+128.742257504" watchObservedRunningTime="2026-03-16 00:08:20.474712558 +0000 UTC m=+128.744044848" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.495900 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:20 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:20 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:20 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.495960 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.505641 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" podStartSLOduration=54.505629023 podStartE2EDuration="54.505629023s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.505344805 +0000 UTC m=+128.774677095" watchObservedRunningTime="2026-03-16 00:08:20.505629023 +0000 UTC m=+128.774961313" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.511362 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.511828 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.011810528 +0000 UTC m=+129.281142818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.597423 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qsvvx" podStartSLOduration=54.597399791 podStartE2EDuration="54.597399791s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.561130135 +0000 UTC m=+128.830462425" watchObservedRunningTime="2026-03-16 00:08:20.597399791 +0000 UTC m=+128.866732081" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.597860 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" podStartSLOduration=53.597855574 podStartE2EDuration="53.597855574s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.59736541 +0000 UTC m=+128.866697700" watchObservedRunningTime="2026-03-16 00:08:20.597855574 +0000 UTC m=+128.867187864" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.613303 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.613625 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.113612176 +0000 UTC m=+129.382944466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.714326 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.714721 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.214707873 +0000 UTC m=+129.484040163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.809395 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-nx86z" podStartSLOduration=53.809374106 podStartE2EDuration="53.809374106s" podCreationTimestamp="2026-03-16 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:20.809211981 +0000 UTC m=+129.078544271" watchObservedRunningTime="2026-03-16 00:08:20.809374106 +0000 UTC m=+129.078706396" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.815969 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.816401 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.316388446 +0000 UTC m=+129.585720736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.850907 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34256: no serving certificate available for the kubelet" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.882370 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rl4gt" Mar 16 00:08:20 crc kubenswrapper[4751]: I0316 00:08:20.917118 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:20 crc kubenswrapper[4751]: E0316 00:08:20.917594 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.417576555 +0000 UTC m=+129.686908845 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.019171 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.019573 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.519554528 +0000 UTC m=+129.788886808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.114524 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8kc6w"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.115471 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.120557 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.120699 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.620674555 +0000 UTC m=+129.890006845 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.120826 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.121162 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.62114572 +0000 UTC m=+129.890478020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.121568 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.129245 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8kc6w"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.174896 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" event={"ID":"74e4f428-25a3-45b9-959e-8a1eeb968cdf","Type":"ContainerStarted","Data":"a530660c91dabc4c110bc02e7996df6ee65257fd9a315abbee9cb59e8e9f257d"} Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.174947 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" event={"ID":"74e4f428-25a3-45b9-959e-8a1eeb968cdf","Type":"ContainerStarted","Data":"aedd0ab7cee7138a46458900350712b0fea670eb508983302381dcc72df36df4"} Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.180376 4751 generic.go:334] "Generic (PLEG): container finished" podID="c2676895-4d0e-4099-9d49-abd723369982" containerID="7ce477121853d95f45d375e8cd1901a4a918606253974a7d9b47ee82aef01a7e" exitCode=0 Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.180578 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" event={"ID":"c2676895-4d0e-4099-9d49-abd723369982","Type":"ContainerDied","Data":"7ce477121853d95f45d375e8cd1901a4a918606253974a7d9b47ee82aef01a7e"} Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.191605 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-69wfh" event={"ID":"b1a71ab9-7855-40cf-9670-86381e49c4bc","Type":"ContainerStarted","Data":"476fa24819e1861f8f83e5d49e97eade57cbe50a7b367d91137b8c82f936d1f6"} Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.207491 4751 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sph8v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.207586 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.208136 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.208141 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-6k5cn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.208222 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6k5cn" podUID="2394c21b-2d69-4698-8e58-d0198ec77558" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.223726 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.221039 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" podStartSLOduration=55.220086002 podStartE2EDuration="55.220086002s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:21.208829415 +0000 UTC m=+129.478161695" watchObservedRunningTime="2026-03-16 00:08:21.220086002 +0000 UTC m=+129.489418292" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.233305 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.238075 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.738038779 +0000 UTC m=+130.007371069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.321565 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mqzbn"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.333074 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6kdp\" (UniqueName: \"kubernetes.io/projected/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-kube-api-access-f6kdp\") pod \"community-operators-8kc6w\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.333243 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.333370 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-utilities\") pod \"community-operators-8kc6w\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.333404 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-catalog-content\") pod \"community-operators-8kc6w\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.340089 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.840069334 +0000 UTC m=+130.109401704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.341059 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.358015 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.360524 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqzbn"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.433823 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.433962 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-catalog-content\") pod \"certified-operators-mqzbn\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.433984 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljkkl\" (UniqueName: \"kubernetes.io/projected/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-kube-api-access-ljkkl\") pod \"certified-operators-mqzbn\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.434011 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6kdp\" (UniqueName: \"kubernetes.io/projected/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-kube-api-access-f6kdp\") pod \"community-operators-8kc6w\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.434056 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-utilities\") pod \"community-operators-8kc6w\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.434075 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-catalog-content\") pod \"community-operators-8kc6w\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.434092 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-utilities\") pod \"certified-operators-mqzbn\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.435189 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:21.935170481 +0000 UTC m=+130.204502771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.435587 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-catalog-content\") pod \"community-operators-8kc6w\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.435681 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-utilities\") pod \"community-operators-8kc6w\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.477876 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6kdp\" (UniqueName: \"kubernetes.io/projected/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-kube-api-access-f6kdp\") pod \"community-operators-8kc6w\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.496281 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:21 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:21 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:21 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.496340 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.513342 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tbr2k"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.514473 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.535090 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-catalog-content\") pod \"certified-operators-mqzbn\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.535143 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljkkl\" (UniqueName: \"kubernetes.io/projected/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-kube-api-access-ljkkl\") pod \"certified-operators-mqzbn\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.535182 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.535228 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-utilities\") pod \"certified-operators-mqzbn\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.535592 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-catalog-content\") pod \"certified-operators-mqzbn\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.536177 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.036160214 +0000 UTC m=+130.305492504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.540473 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-utilities\") pod \"certified-operators-mqzbn\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.547332 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tbr2k"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.567497 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34270: no serving certificate available for the kubelet" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.588841 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljkkl\" (UniqueName: \"kubernetes.io/projected/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-kube-api-access-ljkkl\") pod \"certified-operators-mqzbn\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.618426 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mrm4x"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.636133 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.636382 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-catalog-content\") pod \"community-operators-tbr2k\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.636428 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlr45\" (UniqueName: \"kubernetes.io/projected/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-kube-api-access-rlr45\") pod \"community-operators-tbr2k\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.636453 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-utilities\") pod \"community-operators-tbr2k\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.636580 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.13656086 +0000 UTC m=+130.405893150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.681474 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.682011 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" podUID="5c2eff22-d730-456b-870d-4138d9354577" containerName="route-controller-manager" containerID="cri-o://cc11ef3da5a2375d5ec5d6fe7bc35382471a765532ece0e87f87e9afb820c437" gracePeriod=30 Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.686223 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.687648 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.687995 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.691260 4751 patch_prober.go:28] interesting pod/apiserver-76f77b778f-9mrng container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.691315 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" podUID="74e4f428-25a3-45b9-959e-8a1eeb968cdf" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.718521 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-chg6k"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.719562 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.736043 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.740008 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-catalog-content\") pod \"community-operators-tbr2k\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.740552 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlr45\" (UniqueName: \"kubernetes.io/projected/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-kube-api-access-rlr45\") pod \"community-operators-tbr2k\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.740588 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-utilities\") pod \"community-operators-tbr2k\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.740659 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.740930 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.240918844 +0000 UTC m=+130.510251134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.740489 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-catalog-content\") pod \"community-operators-tbr2k\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.741566 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-utilities\") pod \"community-operators-tbr2k\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.751906 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-chg6k"] Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.786732 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlr45\" (UniqueName: \"kubernetes.io/projected/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-kube-api-access-rlr45\") pod \"community-operators-tbr2k\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.834984 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.841342 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.841656 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-catalog-content\") pod \"certified-operators-chg6k\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.841838 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.341810265 +0000 UTC m=+130.611142555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.842054 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-utilities\") pod \"certified-operators-chg6k\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.842167 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.842220 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d87fq\" (UniqueName: \"kubernetes.io/projected/163b6298-4cca-4bea-9272-124555c2b0f4-kube-api-access-d87fq\") pod \"certified-operators-chg6k\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.842664 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.34265219 +0000 UTC m=+130.611984490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.856975 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.947602 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.947804 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.447774317 +0000 UTC m=+130.717106607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.948161 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.948200 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d87fq\" (UniqueName: \"kubernetes.io/projected/163b6298-4cca-4bea-9272-124555c2b0f4-kube-api-access-d87fq\") pod \"certified-operators-chg6k\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.948228 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-catalog-content\") pod \"certified-operators-chg6k\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.948327 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-utilities\") pod \"certified-operators-chg6k\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:21 crc kubenswrapper[4751]: E0316 00:08:21.948677 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.448667914 +0000 UTC m=+130.718000194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.949414 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-catalog-content\") pod \"certified-operators-chg6k\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.953508 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-utilities\") pod \"certified-operators-chg6k\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:21 crc kubenswrapper[4751]: I0316 00:08:21.972561 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d87fq\" (UniqueName: \"kubernetes.io/projected/163b6298-4cca-4bea-9272-124555c2b0f4-kube-api-access-d87fq\") pod \"certified-operators-chg6k\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.051709 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.052139 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.552124201 +0000 UTC m=+130.821456491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.143485 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.155937 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.156330 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.656317811 +0000 UTC m=+130.925650101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.198296 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-69wfh" event={"ID":"b1a71ab9-7855-40cf-9670-86381e49c4bc","Type":"ContainerStarted","Data":"0bbacb47d58ab0b623868a1d9dd101ffea56b57eff939103dcdaf9198657ae97"} Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.199579 4751 generic.go:334] "Generic (PLEG): container finished" podID="5c2eff22-d730-456b-870d-4138d9354577" containerID="cc11ef3da5a2375d5ec5d6fe7bc35382471a765532ece0e87f87e9afb820c437" exitCode=0 Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.200224 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" event={"ID":"5c2eff22-d730-456b-870d-4138d9354577","Type":"ContainerDied","Data":"cc11ef3da5a2375d5ec5d6fe7bc35382471a765532ece0e87f87e9afb820c437"} Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.208446 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.212295 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-88ldb" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.258785 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.273278 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.773244291 +0000 UTC m=+131.042576581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.362021 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.364613 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.864594456 +0000 UTC m=+131.133926746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.429784 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqzbn"] Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.463055 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.463472 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:22.963454276 +0000 UTC m=+131.232786566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.496066 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:22 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:22 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:22 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.496125 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.500471 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8kc6w"] Mar 16 00:08:22 crc kubenswrapper[4751]: W0316 00:08:22.514766 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d8a4b98_9f6d_4bfa_abba_ee21c09f51df.slice/crio-040740a1624631b07ad2604d55932e813ad46029a4833aa4d0afe1e89736843b WatchSource:0}: Error finding container 040740a1624631b07ad2604d55932e813ad46029a4833aa4d0afe1e89736843b: Status 404 returned error can't find the container with id 040740a1624631b07ad2604d55932e813ad46029a4833aa4d0afe1e89736843b Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.564210 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.564632 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.064614215 +0000 UTC m=+131.333946505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.633212 4751 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.659462 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.665178 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.665819 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.165800894 +0000 UTC m=+131.435133184 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.666419 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tbr2k"] Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.767078 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-client-ca\") pod \"5c2eff22-d730-456b-870d-4138d9354577\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.767328 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fs9l\" (UniqueName: \"kubernetes.io/projected/5c2eff22-d730-456b-870d-4138d9354577-kube-api-access-8fs9l\") pod \"5c2eff22-d730-456b-870d-4138d9354577\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.767394 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c2eff22-d730-456b-870d-4138d9354577-serving-cert\") pod \"5c2eff22-d730-456b-870d-4138d9354577\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.767525 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-config\") pod \"5c2eff22-d730-456b-870d-4138d9354577\" (UID: \"5c2eff22-d730-456b-870d-4138d9354577\") " Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.767655 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.767920 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-client-ca" (OuterVolumeSpecName: "client-ca") pod "5c2eff22-d730-456b-870d-4138d9354577" (UID: "5c2eff22-d730-456b-870d-4138d9354577"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.768034 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.268020834 +0000 UTC m=+131.537353124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.772343 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-config" (OuterVolumeSpecName: "config") pod "5c2eff22-d730-456b-870d-4138d9354577" (UID: "5c2eff22-d730-456b-870d-4138d9354577"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.774505 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c2eff22-d730-456b-870d-4138d9354577-kube-api-access-8fs9l" (OuterVolumeSpecName: "kube-api-access-8fs9l") pod "5c2eff22-d730-456b-870d-4138d9354577" (UID: "5c2eff22-d730-456b-870d-4138d9354577"). InnerVolumeSpecName "kube-api-access-8fs9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.775121 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c2eff22-d730-456b-870d-4138d9354577-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5c2eff22-d730-456b-870d-4138d9354577" (UID: "5c2eff22-d730-456b-870d-4138d9354577"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.868431 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.868921 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fs9l\" (UniqueName: \"kubernetes.io/projected/5c2eff22-d730-456b-870d-4138d9354577-kube-api-access-8fs9l\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.869027 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c2eff22-d730-456b-870d-4138d9354577-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.869123 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.869201 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c2eff22-d730-456b-870d-4138d9354577-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.869315 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.369299236 +0000 UTC m=+131.638631526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.894490 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34282: no serving certificate available for the kubelet" Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.912250 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-chg6k"] Mar 16 00:08:22 crc kubenswrapper[4751]: I0316 00:08:22.974082 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:22 crc kubenswrapper[4751]: E0316 00:08:22.974451 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.474438314 +0000 UTC m=+131.743770604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.036407 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:23 crc kubenswrapper[4751]: W0316 00:08:23.037941 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod163b6298_4cca_4bea_9272_124555c2b0f4.slice/crio-2b723c23a50e62d87c006b204cf4704ab5176d57f415f4c92e7e6618ff41685d WatchSource:0}: Error finding container 2b723c23a50e62d87c006b204cf4704ab5176d57f415f4c92e7e6618ff41685d: Status 404 returned error can't find the container with id 2b723c23a50e62d87c006b204cf4704ab5176d57f415f4c92e7e6618ff41685d Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.075077 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:23 crc kubenswrapper[4751]: E0316 00:08:23.075238 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.575216021 +0000 UTC m=+131.844548311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.075355 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:23 crc kubenswrapper[4751]: E0316 00:08:23.075734 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.575721446 +0000 UTC m=+131.845053736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.176842 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76wnr\" (UniqueName: \"kubernetes.io/projected/c2676895-4d0e-4099-9d49-abd723369982-kube-api-access-76wnr\") pod \"c2676895-4d0e-4099-9d49-abd723369982\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.177228 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2676895-4d0e-4099-9d49-abd723369982-config-volume\") pod \"c2676895-4d0e-4099-9d49-abd723369982\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.177309 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2676895-4d0e-4099-9d49-abd723369982-secret-volume\") pod \"c2676895-4d0e-4099-9d49-abd723369982\" (UID: \"c2676895-4d0e-4099-9d49-abd723369982\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.177468 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:23 crc kubenswrapper[4751]: E0316 00:08:23.177984 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.677946097 +0000 UTC m=+131.947278387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.179879 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2676895-4d0e-4099-9d49-abd723369982-config-volume" (OuterVolumeSpecName: "config-volume") pod "c2676895-4d0e-4099-9d49-abd723369982" (UID: "c2676895-4d0e-4099-9d49-abd723369982"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.184163 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2676895-4d0e-4099-9d49-abd723369982-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c2676895-4d0e-4099-9d49-abd723369982" (UID: "c2676895-4d0e-4099-9d49-abd723369982"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.187218 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2676895-4d0e-4099-9d49-abd723369982-kube-api-access-76wnr" (OuterVolumeSpecName: "kube-api-access-76wnr") pod "c2676895-4d0e-4099-9d49-abd723369982" (UID: "c2676895-4d0e-4099-9d49-abd723369982"). InnerVolumeSpecName "kube-api-access-76wnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.227393 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.228719 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz" event={"ID":"5c2eff22-d730-456b-870d-4138d9354577","Type":"ContainerDied","Data":"1a783ea48f6fb8e6a6b842fa7de6aaa0eb69c5cd0608f9d72bf2f0a2cc8b79c9"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.228783 4751 scope.go:117] "RemoveContainer" containerID="cc11ef3da5a2375d5ec5d6fe7bc35382471a765532ece0e87f87e9afb820c437" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.232725 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.232847 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560320-ncm2w" event={"ID":"c2676895-4d0e-4099-9d49-abd723369982","Type":"ContainerDied","Data":"18da8a0a2bd065aca2ba6dd35af1ea1cc81e950e018d28cd69617a536fcf7ce1"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.232880 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18da8a0a2bd065aca2ba6dd35af1ea1cc81e950e018d28cd69617a536fcf7ce1" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.256456 4751 generic.go:334] "Generic (PLEG): container finished" podID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerID="65ea2db385173b1f397b1d71582271814011a09be369aa8bf4d849ab3b8be80e" exitCode=0 Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.260251 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kc6w" event={"ID":"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df","Type":"ContainerDied","Data":"65ea2db385173b1f397b1d71582271814011a09be369aa8bf4d849ab3b8be80e"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.260331 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kc6w" event={"ID":"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df","Type":"ContainerStarted","Data":"040740a1624631b07ad2604d55932e813ad46029a4833aa4d0afe1e89736843b"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.280352 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.280459 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2676895-4d0e-4099-9d49-abd723369982-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.280474 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76wnr\" (UniqueName: \"kubernetes.io/projected/c2676895-4d0e-4099-9d49-abd723369982-kube-api-access-76wnr\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.280485 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2676895-4d0e-4099-9d49-abd723369982-config-volume\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:23 crc kubenswrapper[4751]: E0316 00:08:23.280759 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.780745244 +0000 UTC m=+132.050077534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.284461 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz"] Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.286932 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wvjtz"] Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.295507 4751 generic.go:334] "Generic (PLEG): container finished" podID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerID="2609e29cdc6919c3ca133323fb2abfbeca46e9dbdb36bac936940f849504b789" exitCode=0 Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.295619 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqzbn" event={"ID":"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c","Type":"ContainerDied","Data":"2609e29cdc6919c3ca133323fb2abfbeca46e9dbdb36bac936940f849504b789"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.295655 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqzbn" event={"ID":"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c","Type":"ContainerStarted","Data":"3588e02fe399b6cf2ef41a37dd4b437bb0fd052f6a59909d9ca73efa576c3923"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.312577 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pc9vc"] Mar 16 00:08:23 crc kubenswrapper[4751]: E0316 00:08:23.312830 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c2eff22-d730-456b-870d-4138d9354577" containerName="route-controller-manager" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.312842 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c2eff22-d730-456b-870d-4138d9354577" containerName="route-controller-manager" Mar 16 00:08:23 crc kubenswrapper[4751]: E0316 00:08:23.312861 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2676895-4d0e-4099-9d49-abd723369982" containerName="collect-profiles" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.312867 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2676895-4d0e-4099-9d49-abd723369982" containerName="collect-profiles" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.312979 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c2eff22-d730-456b-870d-4138d9354577" containerName="route-controller-manager" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.312998 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2676895-4d0e-4099-9d49-abd723369982" containerName="collect-profiles" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.315468 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.317709 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chg6k" event={"ID":"163b6298-4cca-4bea-9272-124555c2b0f4","Type":"ContainerStarted","Data":"2b723c23a50e62d87c006b204cf4704ab5176d57f415f4c92e7e6618ff41685d"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.318060 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.339625 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc9vc"] Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.340336 4751 generic.go:334] "Generic (PLEG): container finished" podID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerID="4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa" exitCode=0 Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.340429 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbr2k" event={"ID":"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8","Type":"ContainerDied","Data":"4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.340476 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbr2k" event={"ID":"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8","Type":"ContainerStarted","Data":"fd718e5457525a1d4645330b108810565fed74d8324c17ea0aba899211525a35"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.354455 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-69wfh" event={"ID":"b1a71ab9-7855-40cf-9670-86381e49c4bc","Type":"ContainerStarted","Data":"f6ababb0ad13396a4f3666bf4fcc595945bcb221b848d9f63b2a33f0013430b2"} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.354843 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" podUID="a8fdc932-6670-45f5-9f69-c15eaae40bc4" containerName="controller-manager" containerID="cri-o://dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b" gracePeriod=30 Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.386584 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.386844 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v8h2\" (UniqueName: \"kubernetes.io/projected/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-kube-api-access-8v8h2\") pod \"redhat-marketplace-pc9vc\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.386885 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-utilities\") pod \"redhat-marketplace-pc9vc\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.386908 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-catalog-content\") pod \"redhat-marketplace-pc9vc\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: E0316 00:08:23.387143 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.887094668 +0000 UTC m=+132.156426958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.488218 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v8h2\" (UniqueName: \"kubernetes.io/projected/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-kube-api-access-8v8h2\") pod \"redhat-marketplace-pc9vc\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.488435 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.488476 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-utilities\") pod \"redhat-marketplace-pc9vc\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.488524 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-catalog-content\") pod \"redhat-marketplace-pc9vc\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: E0316 00:08:23.491172 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-16 00:08:23.991159244 +0000 UTC m=+132.260491534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zd5fn" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.491859 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-utilities\") pod \"redhat-marketplace-pc9vc\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.494445 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-catalog-content\") pod \"redhat-marketplace-pc9vc\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.500802 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:23 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:23 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:23 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.501049 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.520571 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v8h2\" (UniqueName: \"kubernetes.io/projected/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-kube-api-access-8v8h2\") pod \"redhat-marketplace-pc9vc\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.560335 4751 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-16T00:08:22.633243749Z","Handler":null,"Name":""} Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.573417 4751 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.573480 4751 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.589926 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.593587 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.670126 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.670824 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.673427 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.674352 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.675093 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.681227 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.691184 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.695242 4751 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.695273 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.712436 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nwz46"] Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.715222 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.716057 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwz46"] Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.747114 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zd5fn\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.792828 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3570077-5371-499e-b664-7cc4b672a7b0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e3570077-5371-499e-b664-7cc4b672a7b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.792873 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxght\" (UniqueName: \"kubernetes.io/projected/bf3ab855-2797-46d2-b42b-e8a43ea455b5-kube-api-access-mxght\") pod \"redhat-marketplace-nwz46\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.792892 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3570077-5371-499e-b664-7cc4b672a7b0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e3570077-5371-499e-b664-7cc4b672a7b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.792931 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-utilities\") pod \"redhat-marketplace-nwz46\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.792962 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-catalog-content\") pod \"redhat-marketplace-nwz46\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.797833 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.887563 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.894171 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895272 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-proxy-ca-bundles\") pod \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895337 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnqqr\" (UniqueName: \"kubernetes.io/projected/a8fdc932-6670-45f5-9f69-c15eaae40bc4-kube-api-access-hnqqr\") pod \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895385 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fdc932-6670-45f5-9f69-c15eaae40bc4-serving-cert\") pod \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895421 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-config\") pod \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895475 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-client-ca\") pod \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\" (UID: \"a8fdc932-6670-45f5-9f69-c15eaae40bc4\") " Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895726 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3570077-5371-499e-b664-7cc4b672a7b0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e3570077-5371-499e-b664-7cc4b672a7b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895761 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxght\" (UniqueName: \"kubernetes.io/projected/bf3ab855-2797-46d2-b42b-e8a43ea455b5-kube-api-access-mxght\") pod \"redhat-marketplace-nwz46\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895778 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3570077-5371-499e-b664-7cc4b672a7b0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e3570077-5371-499e-b664-7cc4b672a7b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895815 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-utilities\") pod \"redhat-marketplace-nwz46\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.895843 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-catalog-content\") pod \"redhat-marketplace-nwz46\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.896276 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-catalog-content\") pod \"redhat-marketplace-nwz46\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.896492 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-utilities\") pod \"redhat-marketplace-nwz46\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.896506 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3570077-5371-499e-b664-7cc4b672a7b0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e3570077-5371-499e-b664-7cc4b672a7b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.896972 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a8fdc932-6670-45f5-9f69-c15eaae40bc4" (UID: "a8fdc932-6670-45f5-9f69-c15eaae40bc4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.897189 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-config" (OuterVolumeSpecName: "config") pod "a8fdc932-6670-45f5-9f69-c15eaae40bc4" (UID: "a8fdc932-6670-45f5-9f69-c15eaae40bc4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.897593 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-client-ca" (OuterVolumeSpecName: "client-ca") pod "a8fdc932-6670-45f5-9f69-c15eaae40bc4" (UID: "a8fdc932-6670-45f5-9f69-c15eaae40bc4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.906756 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8fdc932-6670-45f5-9f69-c15eaae40bc4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a8fdc932-6670-45f5-9f69-c15eaae40bc4" (UID: "a8fdc932-6670-45f5-9f69-c15eaae40bc4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.908367 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8fdc932-6670-45f5-9f69-c15eaae40bc4-kube-api-access-hnqqr" (OuterVolumeSpecName: "kube-api-access-hnqqr") pod "a8fdc932-6670-45f5-9f69-c15eaae40bc4" (UID: "a8fdc932-6670-45f5-9f69-c15eaae40bc4"). InnerVolumeSpecName "kube-api-access-hnqqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.912094 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3570077-5371-499e-b664-7cc4b672a7b0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e3570077-5371-499e-b664-7cc4b672a7b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.916444 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxght\" (UniqueName: \"kubernetes.io/projected/bf3ab855-2797-46d2-b42b-e8a43ea455b5-kube-api-access-mxght\") pod \"redhat-marketplace-nwz46\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.927806 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc9vc"] Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.995384 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.996600 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.996621 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.996630 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fdc932-6670-45f5-9f69-c15eaae40bc4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.996639 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnqqr\" (UniqueName: \"kubernetes.io/projected/a8fdc932-6670-45f5-9f69-c15eaae40bc4-kube-api-access-hnqqr\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:23 crc kubenswrapper[4751]: I0316 00:08:23.996647 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fdc932-6670-45f5-9f69-c15eaae40bc4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.046245 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.122058 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zd5fn"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.226072 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.309220 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tlrgw"] Mar 16 00:08:24 crc kubenswrapper[4751]: E0316 00:08:24.309468 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8fdc932-6670-45f5-9f69-c15eaae40bc4" containerName="controller-manager" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.309482 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8fdc932-6670-45f5-9f69-c15eaae40bc4" containerName="controller-manager" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.309574 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8fdc932-6670-45f5-9f69-c15eaae40bc4" containerName="controller-manager" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.311214 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.317042 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.326574 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwz46"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.338088 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tlrgw"] Mar 16 00:08:24 crc kubenswrapper[4751]: W0316 00:08:24.347719 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf3ab855_2797_46d2_b42b_e8a43ea455b5.slice/crio-3d6e1b14cf44a97e461b30584f6618c2f46d4bf5b49e1c8ee3932eb7c845a7f3 WatchSource:0}: Error finding container 3d6e1b14cf44a97e461b30584f6618c2f46d4bf5b49e1c8ee3932eb7c845a7f3: Status 404 returned error can't find the container with id 3d6e1b14cf44a97e461b30584f6618c2f46d4bf5b49e1c8ee3932eb7c845a7f3 Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.376176 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwz46" event={"ID":"bf3ab855-2797-46d2-b42b-e8a43ea455b5","Type":"ContainerStarted","Data":"3d6e1b14cf44a97e461b30584f6618c2f46d4bf5b49e1c8ee3932eb7c845a7f3"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.378215 4751 generic.go:334] "Generic (PLEG): container finished" podID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerID="16d6ec0a01195cb2eb46262c506491c9603b0788a1b67dd529ffaeb9995ec261" exitCode=0 Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.378267 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc9vc" event={"ID":"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4","Type":"ContainerDied","Data":"16d6ec0a01195cb2eb46262c506491c9603b0788a1b67dd529ffaeb9995ec261"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.378286 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc9vc" event={"ID":"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4","Type":"ContainerStarted","Data":"d9a1781f1af93f7d7bef98eff88870c22b226e91684879ca3247afc9d6a197a4"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.380920 4751 generic.go:334] "Generic (PLEG): container finished" podID="163b6298-4cca-4bea-9272-124555c2b0f4" containerID="345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3" exitCode=0 Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.380990 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chg6k" event={"ID":"163b6298-4cca-4bea-9272-124555c2b0f4","Type":"ContainerDied","Data":"345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.383367 4751 generic.go:334] "Generic (PLEG): container finished" podID="a8fdc932-6670-45f5-9f69-c15eaae40bc4" containerID="dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b" exitCode=0 Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.383447 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.383584 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" event={"ID":"a8fdc932-6670-45f5-9f69-c15eaae40bc4","Type":"ContainerDied","Data":"dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.383661 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mrm4x" event={"ID":"a8fdc932-6670-45f5-9f69-c15eaae40bc4","Type":"ContainerDied","Data":"1cced3cef8665135e290460c0b3834e20fa11d9913decb63fdeb4cb6588d883b"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.383682 4751 scope.go:117] "RemoveContainer" containerID="dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.387642 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e3570077-5371-499e-b664-7cc4b672a7b0","Type":"ContainerStarted","Data":"420455c829cea9e65d5366147b45a58ef06484da777e53df73dd8fbdce8a7fd8"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.396591 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-69wfh" event={"ID":"b1a71ab9-7855-40cf-9670-86381e49c4bc","Type":"ContainerStarted","Data":"880d015883aaab20faf0752a81c911b1945e210cba91688d142009990835f584"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.406637 4751 scope.go:117] "RemoveContainer" containerID="dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b" Mar 16 00:08:24 crc kubenswrapper[4751]: E0316 00:08:24.407977 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b\": container with ID starting with dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b not found: ID does not exist" containerID="dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.408016 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b"} err="failed to get container status \"dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b\": rpc error: code = NotFound desc = could not find container \"dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b\": container with ID starting with dfd406c5af4b9abb1224d4a1fa0e912209efebc1c58e591399d697eff06c733b not found: ID does not exist" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.412371 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-utilities\") pod \"redhat-operators-tlrgw\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.412424 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-catalog-content\") pod \"redhat-operators-tlrgw\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.412466 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prgw4\" (UniqueName: \"kubernetes.io/projected/8f07095b-d178-4ef1-9269-b47e33d71ef3-kube-api-access-prgw4\") pod \"redhat-operators-tlrgw\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.413769 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" event={"ID":"b695c2c9-0b76-4e93-a427-76e157f001fb","Type":"ContainerStarted","Data":"eaf8445d4fafcb92547b8b51320f812a4fc53e42ed2ad3959af9061004359b23"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.413826 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" event={"ID":"b695c2c9-0b76-4e93-a427-76e157f001fb","Type":"ContainerStarted","Data":"483d3baff0187af0e9facbc9c29a244e7a7e1e6bfbace6d8faa206432bbbdad9"} Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.414210 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.427377 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-69wfh" podStartSLOduration=14.427353748 podStartE2EDuration="14.427353748s" podCreationTimestamp="2026-03-16 00:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:24.425307882 +0000 UTC m=+132.694640172" watchObservedRunningTime="2026-03-16 00:08:24.427353748 +0000 UTC m=+132.696686038" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.437706 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mrm4x"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.441465 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mrm4x"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.487475 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" podStartSLOduration=58.487451187 podStartE2EDuration="58.487451187s" podCreationTimestamp="2026-03-16 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:24.484953486 +0000 UTC m=+132.754285776" watchObservedRunningTime="2026-03-16 00:08:24.487451187 +0000 UTC m=+132.756783477" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.493723 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c2eff22-d730-456b-870d-4138d9354577" path="/var/lib/kubelet/pods/5c2eff22-d730-456b-870d-4138d9354577/volumes" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.494481 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.495008 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8fdc932-6670-45f5-9f69-c15eaae40bc4" path="/var/lib/kubelet/pods/a8fdc932-6670-45f5-9f69-c15eaae40bc4/volumes" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.500676 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:24 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:24 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:24 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.500729 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.501466 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-866745c6b5-tm7qz"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.514903 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657448746d-nchpd"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.515362 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657448746d-nchpd"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.515381 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-866745c6b5-tm7qz"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.515624 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prgw4\" (UniqueName: \"kubernetes.io/projected/8f07095b-d178-4ef1-9269-b47e33d71ef3-kube-api-access-prgw4\") pod \"redhat-operators-tlrgw\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.515743 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-utilities\") pod \"redhat-operators-tlrgw\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.515791 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-catalog-content\") pod \"redhat-operators-tlrgw\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.516403 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.517043 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.518446 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-catalog-content\") pod \"redhat-operators-tlrgw\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.519576 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-utilities\") pod \"redhat-operators-tlrgw\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.524062 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.524942 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.525017 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.525280 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.525559 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.525581 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.525835 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.525972 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.526003 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.526083 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.526248 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.526490 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.535723 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.542501 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prgw4\" (UniqueName: \"kubernetes.io/projected/8f07095b-d178-4ef1-9269-b47e33d71ef3-kube-api-access-prgw4\") pod \"redhat-operators-tlrgw\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.617383 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdlbf\" (UniqueName: \"kubernetes.io/projected/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-kube-api-access-hdlbf\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.617459 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-client-ca\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.617487 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-serving-cert\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.617514 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptxqv\" (UniqueName: \"kubernetes.io/projected/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-kube-api-access-ptxqv\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.617583 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-proxy-ca-bundles\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.617602 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-serving-cert\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.617624 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-config\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.617857 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-config\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.617919 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-client-ca\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.650472 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.719422 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdlbf\" (UniqueName: \"kubernetes.io/projected/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-kube-api-access-hdlbf\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.719798 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-client-ca\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.719839 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-serving-cert\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.719864 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptxqv\" (UniqueName: \"kubernetes.io/projected/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-kube-api-access-ptxqv\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.719908 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-proxy-ca-bundles\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.719923 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-serving-cert\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.719958 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-config\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.719997 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-config\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.720025 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-client-ca\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.720673 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-client-ca\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.722037 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-config\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.722719 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-proxy-ca-bundles\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.725787 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-config\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.726787 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-client-ca\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.727727 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-serving-cert\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.729168 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-serving-cert\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.735894 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c6g5f"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.737158 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.786865 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdlbf\" (UniqueName: \"kubernetes.io/projected/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-kube-api-access-hdlbf\") pod \"route-controller-manager-657448746d-nchpd\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.787473 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptxqv\" (UniqueName: \"kubernetes.io/projected/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-kube-api-access-ptxqv\") pod \"controller-manager-866745c6b5-tm7qz\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.802473 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6g5f"] Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.822521 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-catalog-content\") pod \"redhat-operators-c6g5f\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.822666 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-utilities\") pod \"redhat-operators-c6g5f\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.822683 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdpxs\" (UniqueName: \"kubernetes.io/projected/c83465bd-08af-460e-ad52-a874fb037bbb-kube-api-access-zdpxs\") pod \"redhat-operators-c6g5f\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.839266 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.924437 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-utilities\") pod \"redhat-operators-c6g5f\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.924484 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdpxs\" (UniqueName: \"kubernetes.io/projected/c83465bd-08af-460e-ad52-a874fb037bbb-kube-api-access-zdpxs\") pod \"redhat-operators-c6g5f\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.924541 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-catalog-content\") pod \"redhat-operators-c6g5f\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.925303 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-catalog-content\") pod \"redhat-operators-c6g5f\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.925567 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-utilities\") pod \"redhat-operators-c6g5f\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.948034 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdpxs\" (UniqueName: \"kubernetes.io/projected/c83465bd-08af-460e-ad52-a874fb037bbb-kube-api-access-zdpxs\") pod \"redhat-operators-c6g5f\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:24 crc kubenswrapper[4751]: I0316 00:08:24.964576 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.003556 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.108812 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:08:25 crc kubenswrapper[4751]: E0316 00:08:25.159938 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:25 crc kubenswrapper[4751]: E0316 00:08:25.171255 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:25 crc kubenswrapper[4751]: E0316 00:08:25.205290 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:25 crc kubenswrapper[4751]: E0316 00:08:25.205383 4751 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" podUID="9c56ee92-a6f5-46ec-8076-157268b354c7" containerName="kube-multus-additional-cni-plugins" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.226987 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tlrgw"] Mar 16 00:08:25 crc kubenswrapper[4751]: W0316 00:08:25.277143 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f07095b_d178_4ef1_9269_b47e33d71ef3.slice/crio-55454a5f67103e7241a52b787a94b012b1a2f17dda2afcbab9c898cc061e5c08 WatchSource:0}: Error finding container 55454a5f67103e7241a52b787a94b012b1a2f17dda2afcbab9c898cc061e5c08: Status 404 returned error can't find the container with id 55454a5f67103e7241a52b787a94b012b1a2f17dda2afcbab9c898cc061e5c08 Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.328366 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657448746d-nchpd"] Mar 16 00:08:25 crc kubenswrapper[4751]: W0316 00:08:25.373814 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e1985f7_c64e_4b1a_80d0_fa9947ce4a8a.slice/crio-aca6fbad4ea67b9bbff802e566ab20436701540bddf7fee605ef51bcfdea1714 WatchSource:0}: Error finding container aca6fbad4ea67b9bbff802e566ab20436701540bddf7fee605ef51bcfdea1714: Status 404 returned error can't find the container with id aca6fbad4ea67b9bbff802e566ab20436701540bddf7fee605ef51bcfdea1714 Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.446555 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlrgw" event={"ID":"8f07095b-d178-4ef1-9269-b47e33d71ef3","Type":"ContainerStarted","Data":"55454a5f67103e7241a52b787a94b012b1a2f17dda2afcbab9c898cc061e5c08"} Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.475455 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.481854 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6g5f"] Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.485345 4751 generic.go:334] "Generic (PLEG): container finished" podID="e3570077-5371-499e-b664-7cc4b672a7b0" containerID="d70cc0f2cc77db57c3b51a831b0cd74974480be5ea13346bc62abf2d52062a21" exitCode=0 Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.485427 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e3570077-5371-499e-b664-7cc4b672a7b0","Type":"ContainerDied","Data":"d70cc0f2cc77db57c3b51a831b0cd74974480be5ea13346bc62abf2d52062a21"} Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.485521 4751 ???:1] "http: TLS handshake error from 192.168.126.11:34286: no serving certificate available for the kubelet" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.488583 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" event={"ID":"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a","Type":"ContainerStarted","Data":"aca6fbad4ea67b9bbff802e566ab20436701540bddf7fee605ef51bcfdea1714"} Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.496973 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:25 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:25 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:25 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.497029 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.523690 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-866745c6b5-tm7qz"] Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.527069 4751 generic.go:334] "Generic (PLEG): container finished" podID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerID="2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6" exitCode=0 Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.527153 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwz46" event={"ID":"bf3ab855-2797-46d2-b42b-e8a43ea455b5","Type":"ContainerDied","Data":"2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6"} Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.538204 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=0.5381873 podStartE2EDuration="538.1873ms" podCreationTimestamp="2026-03-16 00:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:25.53759028 +0000 UTC m=+133.806922600" watchObservedRunningTime="2026-03-16 00:08:25.5381873 +0000 UTC m=+133.807519590" Mar 16 00:08:25 crc kubenswrapper[4751]: W0316 00:08:25.552563 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc83465bd_08af_460e_ad52_a874fb037bbb.slice/crio-d7cc3cc7da0d2c3f32593a7d1ae675e767d3a81e8f6964eca865e503327d5d4b WatchSource:0}: Error finding container d7cc3cc7da0d2c3f32593a7d1ae675e767d3a81e8f6964eca865e503327d5d4b: Status 404 returned error can't find the container with id d7cc3cc7da0d2c3f32593a7d1ae675e767d3a81e8f6964eca865e503327d5d4b Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.609030 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.609967 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.616156 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.616311 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.619313 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.747038 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.747532 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.849543 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.849631 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.849765 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.893796 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:25 crc kubenswrapper[4751]: I0316 00:08:25.968431 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.326458 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.361908 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.362051 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.362081 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.362128 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.363729 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.371209 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.373031 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.375677 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.490196 4751 patch_prober.go:28] interesting pod/console-f9d7485db-bd5b9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.490276 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-bd5b9" podUID="cfe3c553-b411-4074-b83a-c8eba2518089" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.492635 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.492682 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.492697 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.501672 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:26 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:26 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:26 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.501779 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.579911 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" event={"ID":"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb","Type":"ContainerStarted","Data":"91ac4dc03d25db8be8ab3bb8453f6ce148a5ccb70483d4343ee76e7f04c6670e"} Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.580306 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" event={"ID":"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb","Type":"ContainerStarted","Data":"f258b80a0e55d014515990aaf27f4b46da84a57247e781bdbbdc293c9fe12c18"} Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.580731 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.588031 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.591821 4751 generic.go:334] "Generic (PLEG): container finished" podID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerID="eea43c3512fed11d289e063414da7264c2d03b1b3f04b97c9d8cd6cc975bb6f3" exitCode=0 Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.591928 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlrgw" event={"ID":"8f07095b-d178-4ef1-9269-b47e33d71ef3","Type":"ContainerDied","Data":"eea43c3512fed11d289e063414da7264c2d03b1b3f04b97c9d8cd6cc975bb6f3"} Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.594867 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.597893 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-6k5cn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.597947 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6k5cn" podUID="2394c21b-2d69-4698-8e58-d0198ec77558" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.598172 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-6k5cn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.598236 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6k5cn" podUID="2394c21b-2d69-4698-8e58-d0198ec77558" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.598500 4751 generic.go:334] "Generic (PLEG): container finished" podID="c83465bd-08af-460e-ad52-a874fb037bbb" containerID="8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455" exitCode=0 Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.599333 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6g5f" event={"ID":"c83465bd-08af-460e-ad52-a874fb037bbb","Type":"ContainerDied","Data":"8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455"} Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.599366 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6g5f" event={"ID":"c83465bd-08af-460e-ad52-a874fb037bbb","Type":"ContainerStarted","Data":"d7cc3cc7da0d2c3f32593a7d1ae675e767d3a81e8f6964eca865e503327d5d4b"} Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.602218 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.605297 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" event={"ID":"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a","Type":"ContainerStarted","Data":"240b042e94e39a153a7dbcce475d36046877acedbf425be484a6617b09cd96df"} Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.606211 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.619510 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.637038 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" podStartSLOduration=4.637015394 podStartE2EDuration="4.637015394s" podCreationTimestamp="2026-03-16 00:08:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:26.601734536 +0000 UTC m=+134.871066826" watchObservedRunningTime="2026-03-16 00:08:26.637015394 +0000 UTC m=+134.906347684" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.641238 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.652090 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65","Type":"ContainerStarted","Data":"6d3366e6c9130ce4ea930feea230a6bfa6e7369cb82400c1df6beb4205d448f1"} Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.720576 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.728572 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-9mrng" Mar 16 00:08:26 crc kubenswrapper[4751]: I0316 00:08:26.811824 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" podStartSLOduration=4.811803092 podStartE2EDuration="4.811803092s" podCreationTimestamp="2026-03-16 00:08:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:26.735382037 +0000 UTC m=+135.004714327" watchObservedRunningTime="2026-03-16 00:08:26.811803092 +0000 UTC m=+135.081135382" Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.369756 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:27 crc kubenswrapper[4751]: W0316 00:08:27.449944 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-919c5c61c02260a69f6d659d35d3e4fdeabfab7b1812ff4c5cdbc9559544ac5e WatchSource:0}: Error finding container 919c5c61c02260a69f6d659d35d3e4fdeabfab7b1812ff4c5cdbc9559544ac5e: Status 404 returned error can't find the container with id 919c5c61c02260a69f6d659d35d3e4fdeabfab7b1812ff4c5cdbc9559544ac5e Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.494521 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:27 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:27 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:27 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.494571 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.498418 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3570077-5371-499e-b664-7cc4b672a7b0-kubelet-dir\") pod \"e3570077-5371-499e-b664-7cc4b672a7b0\" (UID: \"e3570077-5371-499e-b664-7cc4b672a7b0\") " Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.498621 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3570077-5371-499e-b664-7cc4b672a7b0-kube-api-access\") pod \"e3570077-5371-499e-b664-7cc4b672a7b0\" (UID: \"e3570077-5371-499e-b664-7cc4b672a7b0\") " Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.499973 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3570077-5371-499e-b664-7cc4b672a7b0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e3570077-5371-499e-b664-7cc4b672a7b0" (UID: "e3570077-5371-499e-b664-7cc4b672a7b0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.507510 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3570077-5371-499e-b664-7cc4b672a7b0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e3570077-5371-499e-b664-7cc4b672a7b0" (UID: "e3570077-5371-499e-b664-7cc4b672a7b0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.605718 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3570077-5371-499e-b664-7cc4b672a7b0-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.605753 4751 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e3570077-5371-499e-b664-7cc4b672a7b0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.660569 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9ec692e9bcb1e0cb69a8b325fa96e050ebf12ae62f9141fc7cf7082b289d8541"} Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.665136 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65","Type":"ContainerStarted","Data":"8bcffe31f16c1dd9b19b0b6fb724e53b3765f4e3db0a088ab78a457a5540b9df"} Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.670605 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"919c5c61c02260a69f6d659d35d3e4fdeabfab7b1812ff4c5cdbc9559544ac5e"} Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.673317 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.673320 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e3570077-5371-499e-b664-7cc4b672a7b0","Type":"ContainerDied","Data":"420455c829cea9e65d5366147b45a58ef06484da777e53df73dd8fbdce8a7fd8"} Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.673361 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="420455c829cea9e65d5366147b45a58ef06484da777e53df73dd8fbdce8a7fd8" Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.679264 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3b2ed377a5c9e247b3c14f9325a29703068e5a098a32614470e2d598757a13ae"} Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.687720 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.687698716 podStartE2EDuration="2.687698716s" podCreationTimestamp="2026-03-16 00:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:27.686400754 +0000 UTC m=+135.955733064" watchObservedRunningTime="2026-03-16 00:08:27.687698716 +0000 UTC m=+135.957031006" Mar 16 00:08:27 crc kubenswrapper[4751]: I0316 00:08:27.955444 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:08:28 crc kubenswrapper[4751]: I0316 00:08:28.153304 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6d776" Mar 16 00:08:28 crc kubenswrapper[4751]: I0316 00:08:28.496141 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:28 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:28 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:28 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:28 crc kubenswrapper[4751]: I0316 00:08:28.496201 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:28 crc kubenswrapper[4751]: I0316 00:08:28.755194 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1e525fb0df7493c23d6257f7680cb5bc634be134139fb9b8b979de363a11594c"} Mar 16 00:08:28 crc kubenswrapper[4751]: I0316 00:08:28.765212 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ccea9540315ee1c0407cf5ff350f024d72438b2be2e0c2673cd0254fa61f61d6"} Mar 16 00:08:28 crc kubenswrapper[4751]: I0316 00:08:28.765287 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:08:28 crc kubenswrapper[4751]: I0316 00:08:28.795906 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"cd87149a182355b7f7888400a208d9903ed5dee1a7a1982e21b9c61c84384edc"} Mar 16 00:08:28 crc kubenswrapper[4751]: I0316 00:08:28.806856 4751 generic.go:334] "Generic (PLEG): container finished" podID="468a0cc0-cc9e-4f31-89c0-e90a4de7ce65" containerID="8bcffe31f16c1dd9b19b0b6fb724e53b3765f4e3db0a088ab78a457a5540b9df" exitCode=0 Mar 16 00:08:28 crc kubenswrapper[4751]: I0316 00:08:28.807652 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65","Type":"ContainerDied","Data":"8bcffe31f16c1dd9b19b0b6fb724e53b3765f4e3db0a088ab78a457a5540b9df"} Mar 16 00:08:29 crc kubenswrapper[4751]: I0316 00:08:29.493932 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:29 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:29 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:29 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:29 crc kubenswrapper[4751]: I0316 00:08:29.493993 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:30 crc kubenswrapper[4751]: I0316 00:08:30.495333 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:30 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:30 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:30 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:30 crc kubenswrapper[4751]: I0316 00:08:30.495402 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:30 crc kubenswrapper[4751]: I0316 00:08:30.656566 4751 ???:1] "http: TLS handshake error from 192.168.126.11:47616: no serving certificate available for the kubelet" Mar 16 00:08:30 crc kubenswrapper[4751]: I0316 00:08:30.758923 4751 ???:1] "http: TLS handshake error from 192.168.126.11:47620: no serving certificate available for the kubelet" Mar 16 00:08:31 crc kubenswrapper[4751]: I0316 00:08:31.495066 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:31 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:31 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:31 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:31 crc kubenswrapper[4751]: I0316 00:08:31.495147 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:32 crc kubenswrapper[4751]: I0316 00:08:32.493692 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:32 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:32 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:32 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:32 crc kubenswrapper[4751]: I0316 00:08:32.493777 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:33 crc kubenswrapper[4751]: I0316 00:08:33.494349 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:33 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:33 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:33 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:33 crc kubenswrapper[4751]: I0316 00:08:33.494408 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:34 crc kubenswrapper[4751]: I0316 00:08:34.493899 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:34 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:34 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:34 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:34 crc kubenswrapper[4751]: I0316 00:08:34.494138 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:35 crc kubenswrapper[4751]: E0316 00:08:35.158801 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:35 crc kubenswrapper[4751]: E0316 00:08:35.162278 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:35 crc kubenswrapper[4751]: E0316 00:08:35.166248 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:35 crc kubenswrapper[4751]: E0316 00:08:35.166281 4751 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" podUID="9c56ee92-a6f5-46ec-8076-157268b354c7" containerName="kube-multus-additional-cni-plugins" Mar 16 00:08:35 crc kubenswrapper[4751]: I0316 00:08:35.494470 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:35 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:35 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:35 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:35 crc kubenswrapper[4751]: I0316 00:08:35.494552 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.474969 4751 patch_prober.go:28] interesting pod/console-f9d7485db-bd5b9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.475025 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-bd5b9" podUID="cfe3c553-b411-4074-b83a-c8eba2518089" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.493640 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:36 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:36 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:36 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.493716 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.617214 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6k5cn" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.833707 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.844404 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kube-api-access\") pod \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\" (UID: \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\") " Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.844463 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kubelet-dir\") pod \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\" (UID: \"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65\") " Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.844714 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "468a0cc0-cc9e-4f31-89c0-e90a4de7ce65" (UID: "468a0cc0-cc9e-4f31-89c0-e90a4de7ce65"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.850426 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "468a0cc0-cc9e-4f31-89c0-e90a4de7ce65" (UID: "468a0cc0-cc9e-4f31-89c0-e90a4de7ce65"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.927310 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"468a0cc0-cc9e-4f31-89c0-e90a4de7ce65","Type":"ContainerDied","Data":"6d3366e6c9130ce4ea930feea230a6bfa6e7369cb82400c1df6beb4205d448f1"} Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.927357 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d3366e6c9130ce4ea930feea230a6bfa6e7369cb82400c1df6beb4205d448f1" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.927359 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.945924 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:36 crc kubenswrapper[4751]: I0316 00:08:36.945945 4751 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/468a0cc0-cc9e-4f31-89c0-e90a4de7ce65-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:37 crc kubenswrapper[4751]: I0316 00:08:37.493563 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jpvk6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 16 00:08:37 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Mar 16 00:08:37 crc kubenswrapper[4751]: [+]process-running ok Mar 16 00:08:37 crc kubenswrapper[4751]: healthz check failed Mar 16 00:08:37 crc kubenswrapper[4751]: I0316 00:08:37.494056 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jpvk6" podUID="a792c668-ee31-4bd8-a122-37bf8b48f37f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 16 00:08:38 crc kubenswrapper[4751]: I0316 00:08:38.504655 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:38 crc kubenswrapper[4751]: I0316 00:08:38.514236 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-jpvk6" Mar 16 00:08:40 crc kubenswrapper[4751]: I0316 00:08:40.939749 4751 ???:1] "http: TLS handshake error from 192.168.126.11:57410: no serving certificate available for the kubelet" Mar 16 00:08:41 crc kubenswrapper[4751]: I0316 00:08:41.311860 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-866745c6b5-tm7qz"] Mar 16 00:08:41 crc kubenswrapper[4751]: I0316 00:08:41.312168 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" podUID="6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" containerName="controller-manager" containerID="cri-o://91ac4dc03d25db8be8ab3bb8453f6ce148a5ccb70483d4343ee76e7f04c6670e" gracePeriod=30 Mar 16 00:08:41 crc kubenswrapper[4751]: I0316 00:08:41.327274 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657448746d-nchpd"] Mar 16 00:08:41 crc kubenswrapper[4751]: I0316 00:08:41.327546 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" podUID="6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" containerName="route-controller-manager" containerID="cri-o://240b042e94e39a153a7dbcce475d36046877acedbf425be484a6617b09cd96df" gracePeriod=30 Mar 16 00:08:41 crc kubenswrapper[4751]: I0316 00:08:41.964583 4751 generic.go:334] "Generic (PLEG): container finished" podID="6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" containerID="91ac4dc03d25db8be8ab3bb8453f6ce148a5ccb70483d4343ee76e7f04c6670e" exitCode=0 Mar 16 00:08:41 crc kubenswrapper[4751]: I0316 00:08:41.964683 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" event={"ID":"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb","Type":"ContainerDied","Data":"91ac4dc03d25db8be8ab3bb8453f6ce148a5ccb70483d4343ee76e7f04c6670e"} Mar 16 00:08:41 crc kubenswrapper[4751]: I0316 00:08:41.967290 4751 generic.go:334] "Generic (PLEG): container finished" podID="6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" containerID="240b042e94e39a153a7dbcce475d36046877acedbf425be484a6617b09cd96df" exitCode=0 Mar 16 00:08:41 crc kubenswrapper[4751]: I0316 00:08:41.967324 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" event={"ID":"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a","Type":"ContainerDied","Data":"240b042e94e39a153a7dbcce475d36046877acedbf425be484a6617b09cd96df"} Mar 16 00:08:42 crc kubenswrapper[4751]: I0316 00:08:42.469863 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 16 00:08:42 crc kubenswrapper[4751]: E0316 00:08:42.806792 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 16 00:08:42 crc kubenswrapper[4751]: E0316 00:08:42.806976 4751 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 16 00:08:42 crc kubenswrapper[4751]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 16 00:08:42 crc kubenswrapper[4751]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rprdx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29560328-lt56z_openshift-infra(c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 16 00:08:42 crc kubenswrapper[4751]: > logger="UnhandledError" Mar 16 00:08:42 crc kubenswrapper[4751]: E0316 00:08:42.808233 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29560328-lt56z" podUID="c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05" Mar 16 00:08:42 crc kubenswrapper[4751]: E0316 00:08:42.975085 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29560328-lt56z" podUID="c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05" Mar 16 00:08:43 crc kubenswrapper[4751]: I0316 00:08:43.009028 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=1.009008234 podStartE2EDuration="1.009008234s" podCreationTimestamp="2026-03-16 00:08:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:43.003892489 +0000 UTC m=+151.273224779" watchObservedRunningTime="2026-03-16 00:08:43.009008234 +0000 UTC m=+151.278340524" Mar 16 00:08:43 crc kubenswrapper[4751]: I0316 00:08:43.900044 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:08:44 crc kubenswrapper[4751]: I0316 00:08:44.473423 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 16 00:08:44 crc kubenswrapper[4751]: I0316 00:08:44.989488 4751 generic.go:334] "Generic (PLEG): container finished" podID="90685c8c-a90e-4821-9e19-40a677b71265" containerID="d7c7bd7974f3e6c328436d2e50ae6f62328e5c16c85ffcbc211729a42c91093e" exitCode=0 Mar 16 00:08:44 crc kubenswrapper[4751]: I0316 00:08:44.989666 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29560320-xlv4n" event={"ID":"90685c8c-a90e-4821-9e19-40a677b71265","Type":"ContainerDied","Data":"d7c7bd7974f3e6c328436d2e50ae6f62328e5c16c85ffcbc211729a42c91093e"} Mar 16 00:08:45 crc kubenswrapper[4751]: I0316 00:08:45.020941 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.020920512 podStartE2EDuration="1.020920512s" podCreationTimestamp="2026-03-16 00:08:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:45.020681484 +0000 UTC m=+153.290013774" watchObservedRunningTime="2026-03-16 00:08:45.020920512 +0000 UTC m=+153.290252802" Mar 16 00:08:45 crc kubenswrapper[4751]: E0316 00:08:45.164335 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:45 crc kubenswrapper[4751]: E0316 00:08:45.165941 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:45 crc kubenswrapper[4751]: E0316 00:08:45.166946 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:45 crc kubenswrapper[4751]: E0316 00:08:45.166978 4751 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" podUID="9c56ee92-a6f5-46ec-8076-157268b354c7" containerName="kube-multus-additional-cni-plugins" Mar 16 00:08:45 crc kubenswrapper[4751]: I0316 00:08:45.965697 4751 patch_prober.go:28] interesting pod/controller-manager-866745c6b5-tm7qz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 16 00:08:45 crc kubenswrapper[4751]: I0316 00:08:45.965817 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" podUID="6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 16 00:08:46 crc kubenswrapper[4751]: I0316 00:08:46.008434 4751 patch_prober.go:28] interesting pod/route-controller-manager-657448746d-nchpd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 16 00:08:46 crc kubenswrapper[4751]: I0316 00:08:46.008533 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" podUID="6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 16 00:08:46 crc kubenswrapper[4751]: I0316 00:08:46.490851 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:46 crc kubenswrapper[4751]: I0316 00:08:46.497937 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-bd5b9" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.220397 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.236420 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.243187 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.260535 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5fd5869d66-87xgz"] Mar 16 00:08:47 crc kubenswrapper[4751]: E0316 00:08:47.260795 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3570077-5371-499e-b664-7cc4b672a7b0" containerName="pruner" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.260808 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3570077-5371-499e-b664-7cc4b672a7b0" containerName="pruner" Mar 16 00:08:47 crc kubenswrapper[4751]: E0316 00:08:47.260822 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" containerName="route-controller-manager" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.260865 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" containerName="route-controller-manager" Mar 16 00:08:47 crc kubenswrapper[4751]: E0316 00:08:47.260875 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="468a0cc0-cc9e-4f31-89c0-e90a4de7ce65" containerName="pruner" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.260882 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="468a0cc0-cc9e-4f31-89c0-e90a4de7ce65" containerName="pruner" Mar 16 00:08:47 crc kubenswrapper[4751]: E0316 00:08:47.260891 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" containerName="controller-manager" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.260899 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" containerName="controller-manager" Mar 16 00:08:47 crc kubenswrapper[4751]: E0316 00:08:47.260912 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90685c8c-a90e-4821-9e19-40a677b71265" containerName="image-pruner" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.260920 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="90685c8c-a90e-4821-9e19-40a677b71265" containerName="image-pruner" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.261030 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" containerName="route-controller-manager" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.261043 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="468a0cc0-cc9e-4f31-89c0-e90a4de7ce65" containerName="pruner" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.261056 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3570077-5371-499e-b664-7cc4b672a7b0" containerName="pruner" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.261071 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" containerName="controller-manager" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.261079 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="90685c8c-a90e-4821-9e19-40a677b71265" containerName="image-pruner" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.261502 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.304237 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5fd5869d66-87xgz"] Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.346724 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-proxy-ca-bundles\") pod \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347161 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-config\") pod \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347203 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-serving-cert\") pod \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347226 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdlbf\" (UniqueName: \"kubernetes.io/projected/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-kube-api-access-hdlbf\") pod \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347267 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-client-ca\") pod \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347298 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-serving-cert\") pod \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347320 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k42kp\" (UniqueName: \"kubernetes.io/projected/90685c8c-a90e-4821-9e19-40a677b71265-kube-api-access-k42kp\") pod \"90685c8c-a90e-4821-9e19-40a677b71265\" (UID: \"90685c8c-a90e-4821-9e19-40a677b71265\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347343 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-config\") pod \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347381 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/90685c8c-a90e-4821-9e19-40a677b71265-serviceca\") pod \"90685c8c-a90e-4821-9e19-40a677b71265\" (UID: \"90685c8c-a90e-4821-9e19-40a677b71265\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347440 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-client-ca\") pod \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\" (UID: \"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.347464 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptxqv\" (UniqueName: \"kubernetes.io/projected/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-kube-api-access-ptxqv\") pod \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\" (UID: \"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb\") " Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.349418 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-client-ca" (OuterVolumeSpecName: "client-ca") pod "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" (UID: "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.349993 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" (UID: "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.350227 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-config" (OuterVolumeSpecName: "config") pod "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" (UID: "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.350220 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90685c8c-a90e-4821-9e19-40a677b71265-serviceca" (OuterVolumeSpecName: "serviceca") pod "90685c8c-a90e-4821-9e19-40a677b71265" (UID: "90685c8c-a90e-4821-9e19-40a677b71265"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.350586 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-config" (OuterVolumeSpecName: "config") pod "6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" (UID: "6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.351895 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-client-ca" (OuterVolumeSpecName: "client-ca") pod "6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" (UID: "6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.354562 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-kube-api-access-ptxqv" (OuterVolumeSpecName: "kube-api-access-ptxqv") pod "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" (UID: "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb"). InnerVolumeSpecName "kube-api-access-ptxqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.354588 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" (UID: "6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.354620 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-kube-api-access-hdlbf" (OuterVolumeSpecName: "kube-api-access-hdlbf") pod "6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" (UID: "6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a"). InnerVolumeSpecName "kube-api-access-hdlbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.357766 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90685c8c-a90e-4821-9e19-40a677b71265-kube-api-access-k42kp" (OuterVolumeSpecName: "kube-api-access-k42kp") pod "90685c8c-a90e-4821-9e19-40a677b71265" (UID: "90685c8c-a90e-4821-9e19-40a677b71265"). InnerVolumeSpecName "kube-api-access-k42kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.377413 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" (UID: "6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.449576 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-proxy-ca-bundles\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.449653 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-config\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.449673 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d85dc5-94f2-40f7-9b75-bfc9464f3084-serving-cert\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.449694 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-client-ca\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.449988 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4ptj\" (UniqueName: \"kubernetes.io/projected/02d85dc5-94f2-40f7-9b75-bfc9464f3084-kube-api-access-c4ptj\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450271 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450292 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450306 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k42kp\" (UniqueName: \"kubernetes.io/projected/90685c8c-a90e-4821-9e19-40a677b71265-kube-api-access-k42kp\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450321 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450333 4751 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/90685c8c-a90e-4821-9e19-40a677b71265-serviceca\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450346 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450358 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptxqv\" (UniqueName: \"kubernetes.io/projected/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-kube-api-access-ptxqv\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450371 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450384 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450395 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.450406 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdlbf\" (UniqueName: \"kubernetes.io/projected/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a-kube-api-access-hdlbf\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.551831 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-proxy-ca-bundles\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.551902 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-config\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.551930 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d85dc5-94f2-40f7-9b75-bfc9464f3084-serving-cert\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.551949 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-client-ca\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.551994 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4ptj\" (UniqueName: \"kubernetes.io/projected/02d85dc5-94f2-40f7-9b75-bfc9464f3084-kube-api-access-c4ptj\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.553417 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-client-ca\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.553524 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-proxy-ca-bundles\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.553580 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-config\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.560212 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d85dc5-94f2-40f7-9b75-bfc9464f3084-serving-cert\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.567719 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4ptj\" (UniqueName: \"kubernetes.io/projected/02d85dc5-94f2-40f7-9b75-bfc9464f3084-kube-api-access-c4ptj\") pod \"controller-manager-5fd5869d66-87xgz\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:47 crc kubenswrapper[4751]: I0316 00:08:47.619742 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.010722 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" event={"ID":"6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb","Type":"ContainerDied","Data":"f258b80a0e55d014515990aaf27f4b46da84a57247e781bdbbdc293c9fe12c18"} Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.010792 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-866745c6b5-tm7qz" Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.011037 4751 scope.go:117] "RemoveContainer" containerID="91ac4dc03d25db8be8ab3bb8453f6ce148a5ccb70483d4343ee76e7f04c6670e" Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.013313 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" event={"ID":"6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a","Type":"ContainerDied","Data":"aca6fbad4ea67b9bbff802e566ab20436701540bddf7fee605ef51bcfdea1714"} Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.013432 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-657448746d-nchpd" Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.015047 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29560320-xlv4n" event={"ID":"90685c8c-a90e-4821-9e19-40a677b71265","Type":"ContainerDied","Data":"125254adf9db91e886bfc5390613cfe18d8ffd0c40d1ce46fe71928f22fedce1"} Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.015076 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="125254adf9db91e886bfc5390613cfe18d8ffd0c40d1ce46fe71928f22fedce1" Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.015138 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29560320-xlv4n" Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.040806 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-866745c6b5-tm7qz"] Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.044649 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-866745c6b5-tm7qz"] Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.060323 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657448746d-nchpd"] Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.064776 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-657448746d-nchpd"] Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.478341 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb" path="/var/lib/kubelet/pods/6bb49a3b-f46e-4bd8-8e5c-a71ebaa509fb/volumes" Mar 16 00:08:48 crc kubenswrapper[4751]: I0316 00:08:48.479168 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a" path="/var/lib/kubelet/pods/6e1985f7-c64e-4b1a-80d0-fa9947ce4a8a/volumes" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.494405 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm"] Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.495128 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.497223 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.498279 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.498551 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.498872 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.498884 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.499074 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.504760 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm"] Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.580693 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkfhg\" (UniqueName: \"kubernetes.io/projected/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-kube-api-access-jkfhg\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.580756 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-client-ca\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.580808 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-serving-cert\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.580832 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-config\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.681618 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkfhg\" (UniqueName: \"kubernetes.io/projected/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-kube-api-access-jkfhg\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.681670 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-client-ca\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.681702 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-serving-cert\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.681723 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-config\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.682929 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-config\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.683778 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-client-ca\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.689626 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-serving-cert\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.696881 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkfhg\" (UniqueName: \"kubernetes.io/projected/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-kube-api-access-jkfhg\") pod \"route-controller-manager-f7bbd968d-9s4hm\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:49 crc kubenswrapper[4751]: I0316 00:08:49.867285 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:51 crc kubenswrapper[4751]: I0316 00:08:51.033902 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-hz7xk_9c56ee92-a6f5-46ec-8076-157268b354c7/kube-multus-additional-cni-plugins/0.log" Mar 16 00:08:51 crc kubenswrapper[4751]: I0316 00:08:51.034202 4751 generic.go:334] "Generic (PLEG): container finished" podID="9c56ee92-a6f5-46ec-8076-157268b354c7" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" exitCode=137 Mar 16 00:08:51 crc kubenswrapper[4751]: I0316 00:08:51.034246 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" event={"ID":"9c56ee92-a6f5-46ec-8076-157268b354c7","Type":"ContainerDied","Data":"80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413"} Mar 16 00:08:52 crc kubenswrapper[4751]: E0316 00:08:52.870171 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 16 00:08:52 crc kubenswrapper[4751]: E0316 00:08:52.870901 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d87fq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-chg6k_openshift-marketplace(163b6298-4cca-4bea-9272-124555c2b0f4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 16 00:08:52 crc kubenswrapper[4751]: E0316 00:08:52.872216 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-chg6k" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" Mar 16 00:08:54 crc kubenswrapper[4751]: E0316 00:08:54.345951 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-chg6k" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" Mar 16 00:08:54 crc kubenswrapper[4751]: E0316 00:08:54.438838 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 16 00:08:54 crc kubenswrapper[4751]: E0316 00:08:54.439007 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8v8h2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pc9vc_openshift-marketplace(5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 16 00:08:54 crc kubenswrapper[4751]: E0316 00:08:54.440291 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pc9vc" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" Mar 16 00:08:54 crc kubenswrapper[4751]: E0316 00:08:54.498986 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 16 00:08:54 crc kubenswrapper[4751]: E0316 00:08:54.499463 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ljkkl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-mqzbn_openshift-marketplace(8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 16 00:08:54 crc kubenswrapper[4751]: E0316 00:08:54.500939 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-mqzbn" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" Mar 16 00:08:55 crc kubenswrapper[4751]: I0316 00:08:55.004317 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sfnlp"] Mar 16 00:08:55 crc kubenswrapper[4751]: E0316 00:08:55.156417 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413 is running failed: container process not found" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:55 crc kubenswrapper[4751]: E0316 00:08:55.156822 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413 is running failed: container process not found" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:55 crc kubenswrapper[4751]: E0316 00:08:55.157358 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413 is running failed: container process not found" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 16 00:08:55 crc kubenswrapper[4751]: E0316 00:08:55.157433 4751 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413 is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" podUID="9c56ee92-a6f5-46ec-8076-157268b354c7" containerName="kube-multus-additional-cni-plugins" Mar 16 00:08:55 crc kubenswrapper[4751]: E0316 00:08:55.916167 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-mqzbn" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" Mar 16 00:08:55 crc kubenswrapper[4751]: I0316 00:08:55.960603 4751 scope.go:117] "RemoveContainer" containerID="240b042e94e39a153a7dbcce475d36046877acedbf425be484a6617b09cd96df" Mar 16 00:08:55 crc kubenswrapper[4751]: E0316 00:08:55.991226 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 16 00:08:55 crc kubenswrapper[4751]: E0316 00:08:55.991900 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f6kdp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-8kc6w_openshift-marketplace(1d8a4b98-9f6d-4bfa-abba-ee21c09f51df): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 16 00:08:55 crc kubenswrapper[4751]: E0316 00:08:55.993296 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-8kc6w" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.001977 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.002154 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rlr45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tbr2k_openshift-marketplace(f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.003515 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tbr2k" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.026305 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-hz7xk_9c56ee92-a6f5-46ec-8076-157268b354c7/kube-multus-additional-cni-plugins/0.log" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.026369 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.050219 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.050363 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-prgw4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-tlrgw_openshift-marketplace(8f07095b-d178-4ef1-9269-b47e33d71ef3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.051486 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-tlrgw" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.072415 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-hz7xk_9c56ee92-a6f5-46ec-8076-157268b354c7/kube-multus-additional-cni-plugins/0.log" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.072493 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" event={"ID":"9c56ee92-a6f5-46ec-8076-157268b354c7","Type":"ContainerDied","Data":"10ba323e9d8c51e682d7c9ed58e560dcf8554ed1a0a6bc5e06bd607d18b50c79"} Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.072544 4751 scope.go:117] "RemoveContainer" containerID="80628b5d69c3338e73346d908ee66d5bc052050eb8841c0a75f68333f239c413" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.072628 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-hz7xk" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.078825 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tbr2k" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.079081 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-8kc6w" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.084480 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-tlrgw" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.093413 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d8ht\" (UniqueName: \"kubernetes.io/projected/9c56ee92-a6f5-46ec-8076-157268b354c7-kube-api-access-8d8ht\") pod \"9c56ee92-a6f5-46ec-8076-157268b354c7\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.093491 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9c56ee92-a6f5-46ec-8076-157268b354c7-tuning-conf-dir\") pod \"9c56ee92-a6f5-46ec-8076-157268b354c7\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.093579 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9c56ee92-a6f5-46ec-8076-157268b354c7-ready\") pod \"9c56ee92-a6f5-46ec-8076-157268b354c7\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.093654 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9c56ee92-a6f5-46ec-8076-157268b354c7-cni-sysctl-allowlist\") pod \"9c56ee92-a6f5-46ec-8076-157268b354c7\" (UID: \"9c56ee92-a6f5-46ec-8076-157268b354c7\") " Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.095188 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c56ee92-a6f5-46ec-8076-157268b354c7-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "9c56ee92-a6f5-46ec-8076-157268b354c7" (UID: "9c56ee92-a6f5-46ec-8076-157268b354c7"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.095313 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c56ee92-a6f5-46ec-8076-157268b354c7-ready" (OuterVolumeSpecName: "ready") pod "9c56ee92-a6f5-46ec-8076-157268b354c7" (UID: "9c56ee92-a6f5-46ec-8076-157268b354c7"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.094454 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c56ee92-a6f5-46ec-8076-157268b354c7-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "9c56ee92-a6f5-46ec-8076-157268b354c7" (UID: "9c56ee92-a6f5-46ec-8076-157268b354c7"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.101328 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c56ee92-a6f5-46ec-8076-157268b354c7-kube-api-access-8d8ht" (OuterVolumeSpecName: "kube-api-access-8d8ht") pod "9c56ee92-a6f5-46ec-8076-157268b354c7" (UID: "9c56ee92-a6f5-46ec-8076-157268b354c7"). InnerVolumeSpecName "kube-api-access-8d8ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.133471 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.133678 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mxght,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-nwz46_openshift-marketplace(bf3ab855-2797-46d2-b42b-e8a43ea455b5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 16 00:08:56 crc kubenswrapper[4751]: E0316 00:08:56.135178 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-nwz46" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.195438 4751 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9c56ee92-a6f5-46ec-8076-157268b354c7-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.195580 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d8ht\" (UniqueName: \"kubernetes.io/projected/9c56ee92-a6f5-46ec-8076-157268b354c7-kube-api-access-8d8ht\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.195751 4751 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9c56ee92-a6f5-46ec-8076-157268b354c7-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.195800 4751 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9c56ee92-a6f5-46ec-8076-157268b354c7-ready\") on node \"crc\" DevicePath \"\"" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.273294 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm"] Mar 16 00:08:56 crc kubenswrapper[4751]: W0316 00:08:56.278178 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f04fb34_7c70_4fda_97ef_d5ed26fe5a2f.slice/crio-b081ce76400e9cbf3bb4816a2a25db063ee564fd77f2c3ce423cbe08e1f6b5c1 WatchSource:0}: Error finding container b081ce76400e9cbf3bb4816a2a25db063ee564fd77f2c3ce423cbe08e1f6b5c1: Status 404 returned error can't find the container with id b081ce76400e9cbf3bb4816a2a25db063ee564fd77f2c3ce423cbe08e1f6b5c1 Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.378779 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5fd5869d66-87xgz"] Mar 16 00:08:56 crc kubenswrapper[4751]: W0316 00:08:56.391990 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02d85dc5_94f2_40f7_9b75_bfc9464f3084.slice/crio-d45df2806aef3f3fcf793ea9b6ca23915a579471448d6a313c18b850f80f5b02 WatchSource:0}: Error finding container d45df2806aef3f3fcf793ea9b6ca23915a579471448d6a313c18b850f80f5b02: Status 404 returned error can't find the container with id d45df2806aef3f3fcf793ea9b6ca23915a579471448d6a313c18b850f80f5b02 Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.421899 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-hz7xk"] Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.427762 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-hz7xk"] Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.503158 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c56ee92-a6f5-46ec-8076-157268b354c7" path="/var/lib/kubelet/pods/9c56ee92-a6f5-46ec-8076-157268b354c7/volumes" Mar 16 00:08:56 crc kubenswrapper[4751]: I0316 00:08:56.835004 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xp482" Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.082447 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" event={"ID":"02d85dc5-94f2-40f7-9b75-bfc9464f3084","Type":"ContainerStarted","Data":"0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952"} Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.082489 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" event={"ID":"02d85dc5-94f2-40f7-9b75-bfc9464f3084","Type":"ContainerStarted","Data":"d45df2806aef3f3fcf793ea9b6ca23915a579471448d6a313c18b850f80f5b02"} Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.083604 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.085055 4751 generic.go:334] "Generic (PLEG): container finished" podID="c83465bd-08af-460e-ad52-a874fb037bbb" containerID="d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb" exitCode=0 Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.085116 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6g5f" event={"ID":"c83465bd-08af-460e-ad52-a874fb037bbb","Type":"ContainerDied","Data":"d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb"} Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.088180 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" event={"ID":"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f","Type":"ContainerStarted","Data":"4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b"} Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.088230 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" event={"ID":"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f","Type":"ContainerStarted","Data":"b081ce76400e9cbf3bb4816a2a25db063ee564fd77f2c3ce423cbe08e1f6b5c1"} Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.088516 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:57 crc kubenswrapper[4751]: E0316 00:08:57.089213 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-nwz46" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.090519 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.093904 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.101287 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" podStartSLOduration=16.101275818 podStartE2EDuration="16.101275818s" podCreationTimestamp="2026-03-16 00:08:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:57.099126959 +0000 UTC m=+165.368459249" watchObservedRunningTime="2026-03-16 00:08:57.101275818 +0000 UTC m=+165.370608108" Mar 16 00:08:57 crc kubenswrapper[4751]: I0316 00:08:57.163718 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" podStartSLOduration=16.163699042 podStartE2EDuration="16.163699042s" podCreationTimestamp="2026-03-16 00:08:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:08:57.139374037 +0000 UTC m=+165.408706327" watchObservedRunningTime="2026-03-16 00:08:57.163699042 +0000 UTC m=+165.433031332" Mar 16 00:08:58 crc kubenswrapper[4751]: I0316 00:08:58.931501 4751 csr.go:261] certificate signing request csr-5dclj is approved, waiting to be issued Mar 16 00:08:58 crc kubenswrapper[4751]: I0316 00:08:58.940594 4751 csr.go:257] certificate signing request csr-5dclj is issued Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.011368 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 16 00:08:59 crc kubenswrapper[4751]: E0316 00:08:59.012135 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c56ee92-a6f5-46ec-8076-157268b354c7" containerName="kube-multus-additional-cni-plugins" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.012159 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c56ee92-a6f5-46ec-8076-157268b354c7" containerName="kube-multus-additional-cni-plugins" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.012303 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c56ee92-a6f5-46ec-8076-157268b354c7" containerName="kube-multus-additional-cni-plugins" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.012824 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.015970 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.016142 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.026230 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.029453 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.029563 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.102598 4751 generic.go:334] "Generic (PLEG): container finished" podID="c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05" containerID="88d600e71538bb890b63c94cdfa3945488e931f89777334bdbf03d3f8e874009" exitCode=0 Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.102721 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560328-lt56z" event={"ID":"c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05","Type":"ContainerDied","Data":"88d600e71538bb890b63c94cdfa3945488e931f89777334bdbf03d3f8e874009"} Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.106373 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6g5f" event={"ID":"c83465bd-08af-460e-ad52-a874fb037bbb","Type":"ContainerStarted","Data":"e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24"} Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.130712 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.130796 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.132665 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.142916 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c6g5f" podStartSLOduration=3.471661807 podStartE2EDuration="35.142889644s" podCreationTimestamp="2026-03-16 00:08:24 +0000 UTC" firstStartedPulling="2026-03-16 00:08:26.618537658 +0000 UTC m=+134.887869948" lastFinishedPulling="2026-03-16 00:08:58.289765455 +0000 UTC m=+166.559097785" observedRunningTime="2026-03-16 00:08:59.142196821 +0000 UTC m=+167.411529141" watchObservedRunningTime="2026-03-16 00:08:59.142889644 +0000 UTC m=+167.412221934" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.160973 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.360932 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.821796 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.942158 4751 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-19 03:55:20.570692327 +0000 UTC Mar 16 00:08:59 crc kubenswrapper[4751]: I0316 00:08:59.942220 4751 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6675h46m20.62847432s for next certificate rotation Mar 16 00:09:00 crc kubenswrapper[4751]: I0316 00:09:00.117678 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25","Type":"ContainerStarted","Data":"744295f37662d928f71ed46d7e356dd3dce36e1a68a69e75a728c8aa09f49136"} Mar 16 00:09:00 crc kubenswrapper[4751]: I0316 00:09:00.453163 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560328-lt56z" Mar 16 00:09:00 crc kubenswrapper[4751]: I0316 00:09:00.553610 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rprdx\" (UniqueName: \"kubernetes.io/projected/c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05-kube-api-access-rprdx\") pod \"c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05\" (UID: \"c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05\") " Mar 16 00:09:00 crc kubenswrapper[4751]: I0316 00:09:00.560445 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05-kube-api-access-rprdx" (OuterVolumeSpecName: "kube-api-access-rprdx") pod "c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05" (UID: "c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05"). InnerVolumeSpecName "kube-api-access-rprdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:00 crc kubenswrapper[4751]: I0316 00:09:00.655003 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rprdx\" (UniqueName: \"kubernetes.io/projected/c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05-kube-api-access-rprdx\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:00 crc kubenswrapper[4751]: I0316 00:09:00.942689 4751 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-25 20:45:20.895150266 +0000 UTC Mar 16 00:09:00 crc kubenswrapper[4751]: I0316 00:09:00.942747 4751 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6836h36m19.952407376s for next certificate rotation Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.124925 4751 generic.go:334] "Generic (PLEG): container finished" podID="22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25" containerID="279de65544e6f5c5502657edcfd5362ff782b60a21148b84575c6f36152a16cc" exitCode=0 Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.125061 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25","Type":"ContainerDied","Data":"279de65544e6f5c5502657edcfd5362ff782b60a21148b84575c6f36152a16cc"} Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.127989 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560328-lt56z" event={"ID":"c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05","Type":"ContainerDied","Data":"4898c76f3dc20881acc5936dd31aa7be9d7710f3784cc9d50d190e8e93058c1d"} Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.128041 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4898c76f3dc20881acc5936dd31aa7be9d7710f3784cc9d50d190e8e93058c1d" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.128141 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560328-lt56z" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.308605 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5fd5869d66-87xgz"] Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.308813 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" podUID="02d85dc5-94f2-40f7-9b75-bfc9464f3084" containerName="controller-manager" containerID="cri-o://0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952" gracePeriod=30 Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.405139 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm"] Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.405340 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" podUID="1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" containerName="route-controller-manager" containerID="cri-o://4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b" gracePeriod=30 Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.879212 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.883738 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.973111 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-serving-cert\") pod \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.973194 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d85dc5-94f2-40f7-9b75-bfc9464f3084-serving-cert\") pod \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.973273 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkfhg\" (UniqueName: \"kubernetes.io/projected/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-kube-api-access-jkfhg\") pod \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.973304 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-config\") pod \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.973321 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-client-ca\") pod \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.973343 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4ptj\" (UniqueName: \"kubernetes.io/projected/02d85dc5-94f2-40f7-9b75-bfc9464f3084-kube-api-access-c4ptj\") pod \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.973366 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-client-ca\") pod \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.973393 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-proxy-ca-bundles\") pod \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\" (UID: \"02d85dc5-94f2-40f7-9b75-bfc9464f3084\") " Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.973412 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-config\") pod \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\" (UID: \"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f\") " Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.974255 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-config" (OuterVolumeSpecName: "config") pod "1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" (UID: "1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.979644 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.979990 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-client-ca" (OuterVolumeSpecName: "client-ca") pod "1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" (UID: "1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.980079 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-client-ca" (OuterVolumeSpecName: "client-ca") pod "02d85dc5-94f2-40f7-9b75-bfc9464f3084" (UID: "02d85dc5-94f2-40f7-9b75-bfc9464f3084"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.980657 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-config" (OuterVolumeSpecName: "config") pod "02d85dc5-94f2-40f7-9b75-bfc9464f3084" (UID: "02d85dc5-94f2-40f7-9b75-bfc9464f3084"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.981163 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "02d85dc5-94f2-40f7-9b75-bfc9464f3084" (UID: "02d85dc5-94f2-40f7-9b75-bfc9464f3084"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.981614 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" (UID: "1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.982628 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-kube-api-access-jkfhg" (OuterVolumeSpecName: "kube-api-access-jkfhg") pod "1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" (UID: "1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f"). InnerVolumeSpecName "kube-api-access-jkfhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.983756 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02d85dc5-94f2-40f7-9b75-bfc9464f3084-kube-api-access-c4ptj" (OuterVolumeSpecName: "kube-api-access-c4ptj") pod "02d85dc5-94f2-40f7-9b75-bfc9464f3084" (UID: "02d85dc5-94f2-40f7-9b75-bfc9464f3084"). InnerVolumeSpecName "kube-api-access-c4ptj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:01 crc kubenswrapper[4751]: I0316 00:09:01.984306 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02d85dc5-94f2-40f7-9b75-bfc9464f3084-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "02d85dc5-94f2-40f7-9b75-bfc9464f3084" (UID: "02d85dc5-94f2-40f7-9b75-bfc9464f3084"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.081502 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d85dc5-94f2-40f7-9b75-bfc9464f3084-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.081564 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkfhg\" (UniqueName: \"kubernetes.io/projected/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-kube-api-access-jkfhg\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.081586 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.081598 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.081612 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4ptj\" (UniqueName: \"kubernetes.io/projected/02d85dc5-94f2-40f7-9b75-bfc9464f3084-kube-api-access-c4ptj\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.081625 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.081638 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/02d85dc5-94f2-40f7-9b75-bfc9464f3084-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.081680 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.136769 4751 generic.go:334] "Generic (PLEG): container finished" podID="02d85dc5-94f2-40f7-9b75-bfc9464f3084" containerID="0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952" exitCode=0 Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.136854 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.136902 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" event={"ID":"02d85dc5-94f2-40f7-9b75-bfc9464f3084","Type":"ContainerDied","Data":"0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952"} Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.136969 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5fd5869d66-87xgz" event={"ID":"02d85dc5-94f2-40f7-9b75-bfc9464f3084","Type":"ContainerDied","Data":"d45df2806aef3f3fcf793ea9b6ca23915a579471448d6a313c18b850f80f5b02"} Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.137002 4751 scope.go:117] "RemoveContainer" containerID="0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.138472 4751 generic.go:334] "Generic (PLEG): container finished" podID="1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" containerID="4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b" exitCode=0 Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.138581 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.138836 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" event={"ID":"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f","Type":"ContainerDied","Data":"4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b"} Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.138927 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm" event={"ID":"1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f","Type":"ContainerDied","Data":"b081ce76400e9cbf3bb4816a2a25db063ee564fd77f2c3ce423cbe08e1f6b5c1"} Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.182080 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm"] Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.188999 4751 scope.go:117] "RemoveContainer" containerID="0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952" Mar 16 00:09:02 crc kubenswrapper[4751]: E0316 00:09:02.189701 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952\": container with ID starting with 0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952 not found: ID does not exist" containerID="0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.189744 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952"} err="failed to get container status \"0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952\": rpc error: code = NotFound desc = could not find container \"0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952\": container with ID starting with 0b83abadb45e5436a7344abc32a818fc60fa29907b921c9a69649aa3513e8952 not found: ID does not exist" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.189774 4751 scope.go:117] "RemoveContainer" containerID="4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.193786 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f7bbd968d-9s4hm"] Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.197789 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5fd5869d66-87xgz"] Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.202316 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5fd5869d66-87xgz"] Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.219296 4751 scope.go:117] "RemoveContainer" containerID="4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b" Mar 16 00:09:02 crc kubenswrapper[4751]: E0316 00:09:02.220000 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b\": container with ID starting with 4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b not found: ID does not exist" containerID="4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.220046 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b"} err="failed to get container status \"4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b\": rpc error: code = NotFound desc = could not find container \"4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b\": container with ID starting with 4e17467dc0477a3db24535fc30b505394694c11f138cd7e9680766c0275e4a3b not found: ID does not exist" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.448048 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.466533 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02d85dc5-94f2-40f7-9b75-bfc9464f3084" path="/var/lib/kubelet/pods/02d85dc5-94f2-40f7-9b75-bfc9464f3084/volumes" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.468892 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" path="/var/lib/kubelet/pods/1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f/volumes" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.484590 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kube-api-access\") pod \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\" (UID: \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\") " Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.484841 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kubelet-dir\") pod \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\" (UID: \"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25\") " Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.484945 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25" (UID: "22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.485221 4751 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.489082 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25" (UID: "22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508088 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-9d557989c-t8b7g"] Mar 16 00:09:02 crc kubenswrapper[4751]: E0316 00:09:02.508408 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" containerName="route-controller-manager" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508430 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" containerName="route-controller-manager" Mar 16 00:09:02 crc kubenswrapper[4751]: E0316 00:09:02.508451 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25" containerName="pruner" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508457 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25" containerName="pruner" Mar 16 00:09:02 crc kubenswrapper[4751]: E0316 00:09:02.508466 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05" containerName="oc" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508472 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05" containerName="oc" Mar 16 00:09:02 crc kubenswrapper[4751]: E0316 00:09:02.508479 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02d85dc5-94f2-40f7-9b75-bfc9464f3084" containerName="controller-manager" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508486 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="02d85dc5-94f2-40f7-9b75-bfc9464f3084" containerName="controller-manager" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508584 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25" containerName="pruner" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508598 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05" containerName="oc" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508606 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="02d85dc5-94f2-40f7-9b75-bfc9464f3084" containerName="controller-manager" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508614 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f04fb34-7c70-4fda-97ef-d5ed26fe5a2f" containerName="route-controller-manager" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.508993 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.515617 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp"] Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.516609 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.518287 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.518611 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.518812 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.519366 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.519371 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.519470 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.521833 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.523244 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.523298 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.523339 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.523473 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.523653 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.523475 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9d557989c-t8b7g"] Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.523967 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.529906 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp"] Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.585951 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce637707-a0df-4f7e-83fc-dadbd0857285-serving-cert\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.585995 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-client-ca\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.586020 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-config\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.586039 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-proxy-ca-bundles\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.586057 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-client-ca\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.586078 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58afe499-1e5d-40ea-aa87-78139be5ff44-serving-cert\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.586094 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-config\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.586131 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcjf8\" (UniqueName: \"kubernetes.io/projected/ce637707-a0df-4f7e-83fc-dadbd0857285-kube-api-access-tcjf8\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.586149 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt6gn\" (UniqueName: \"kubernetes.io/projected/58afe499-1e5d-40ea-aa87-78139be5ff44-kube-api-access-pt6gn\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.586184 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.686707 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-proxy-ca-bundles\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.686765 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-client-ca\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.686805 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58afe499-1e5d-40ea-aa87-78139be5ff44-serving-cert\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.686824 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-config\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.687705 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-client-ca\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.687780 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcjf8\" (UniqueName: \"kubernetes.io/projected/ce637707-a0df-4f7e-83fc-dadbd0857285-kube-api-access-tcjf8\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.687801 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt6gn\" (UniqueName: \"kubernetes.io/projected/58afe499-1e5d-40ea-aa87-78139be5ff44-kube-api-access-pt6gn\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.687849 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce637707-a0df-4f7e-83fc-dadbd0857285-serving-cert\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.687872 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-client-ca\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.687896 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-config\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.688026 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-proxy-ca-bundles\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.688461 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-config\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.688925 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-client-ca\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.690731 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-config\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.691260 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58afe499-1e5d-40ea-aa87-78139be5ff44-serving-cert\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.695833 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce637707-a0df-4f7e-83fc-dadbd0857285-serving-cert\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.706716 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcjf8\" (UniqueName: \"kubernetes.io/projected/ce637707-a0df-4f7e-83fc-dadbd0857285-kube-api-access-tcjf8\") pod \"controller-manager-9d557989c-t8b7g\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.709526 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt6gn\" (UniqueName: \"kubernetes.io/projected/58afe499-1e5d-40ea-aa87-78139be5ff44-kube-api-access-pt6gn\") pod \"route-controller-manager-6c874bf775-cg2cp\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.836258 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:02 crc kubenswrapper[4751]: I0316 00:09:02.842571 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:03 crc kubenswrapper[4751]: I0316 00:09:03.082988 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9d557989c-t8b7g"] Mar 16 00:09:03 crc kubenswrapper[4751]: I0316 00:09:03.104517 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp"] Mar 16 00:09:03 crc kubenswrapper[4751]: W0316 00:09:03.117071 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58afe499_1e5d_40ea_aa87_78139be5ff44.slice/crio-563e5a27107ca64ee3b7e277a84770e3c6dc795ac9bb3868192b0fc4a3e80533 WatchSource:0}: Error finding container 563e5a27107ca64ee3b7e277a84770e3c6dc795ac9bb3868192b0fc4a3e80533: Status 404 returned error can't find the container with id 563e5a27107ca64ee3b7e277a84770e3c6dc795ac9bb3868192b0fc4a3e80533 Mar 16 00:09:03 crc kubenswrapper[4751]: I0316 00:09:03.147383 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"22e3d6c3-80db-4ab3-aa8e-f3a20e19ba25","Type":"ContainerDied","Data":"744295f37662d928f71ed46d7e356dd3dce36e1a68a69e75a728c8aa09f49136"} Mar 16 00:09:03 crc kubenswrapper[4751]: I0316 00:09:03.147426 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="744295f37662d928f71ed46d7e356dd3dce36e1a68a69e75a728c8aa09f49136" Mar 16 00:09:03 crc kubenswrapper[4751]: I0316 00:09:03.147455 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 16 00:09:03 crc kubenswrapper[4751]: I0316 00:09:03.153158 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" event={"ID":"58afe499-1e5d-40ea-aa87-78139be5ff44","Type":"ContainerStarted","Data":"563e5a27107ca64ee3b7e277a84770e3c6dc795ac9bb3868192b0fc4a3e80533"} Mar 16 00:09:03 crc kubenswrapper[4751]: I0316 00:09:03.155183 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" event={"ID":"ce637707-a0df-4f7e-83fc-dadbd0857285","Type":"ContainerStarted","Data":"0743de9b9f39b96eb670cac11c4e829c13dfd4e3ecdcc89c9878a2f195931588"} Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.005521 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.006212 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.007904 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.013192 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.018835 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.120836 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-var-lock\") pod \"installer-9-crc\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.121610 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abaf8528-a999-4343-a42b-66283fd65da0-kube-api-access\") pod \"installer-9-crc\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.121760 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-kubelet-dir\") pod \"installer-9-crc\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.161875 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" event={"ID":"58afe499-1e5d-40ea-aa87-78139be5ff44","Type":"ContainerStarted","Data":"3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670"} Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.163389 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.165454 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" event={"ID":"ce637707-a0df-4f7e-83fc-dadbd0857285","Type":"ContainerStarted","Data":"91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e"} Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.166036 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.171628 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.172651 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.178312 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" podStartSLOduration=3.178284086 podStartE2EDuration="3.178284086s" podCreationTimestamp="2026-03-16 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:09:04.176954953 +0000 UTC m=+172.446287253" watchObservedRunningTime="2026-03-16 00:09:04.178284086 +0000 UTC m=+172.447616376" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.229428 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-var-lock\") pod \"installer-9-crc\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.229516 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abaf8528-a999-4343-a42b-66283fd65da0-kube-api-access\") pod \"installer-9-crc\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.229584 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-kubelet-dir\") pod \"installer-9-crc\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.229659 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-kubelet-dir\") pod \"installer-9-crc\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.231011 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-var-lock\") pod \"installer-9-crc\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.234174 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" podStartSLOduration=3.234164129 podStartE2EDuration="3.234164129s" podCreationTimestamp="2026-03-16 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:09:04.228771365 +0000 UTC m=+172.498103655" watchObservedRunningTime="2026-03-16 00:09:04.234164129 +0000 UTC m=+172.503496419" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.255741 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abaf8528-a999-4343-a42b-66283fd65da0-kube-api-access\") pod \"installer-9-crc\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.325699 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:04 crc kubenswrapper[4751]: I0316 00:09:04.519266 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 16 00:09:05 crc kubenswrapper[4751]: I0316 00:09:05.109356 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:09:05 crc kubenswrapper[4751]: I0316 00:09:05.109677 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:09:05 crc kubenswrapper[4751]: I0316 00:09:05.173014 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"abaf8528-a999-4343-a42b-66283fd65da0","Type":"ContainerStarted","Data":"f598c75bdb47a637122ffd75060f5d76ca629abf1c71116d823ba544c7dff990"} Mar 16 00:09:05 crc kubenswrapper[4751]: I0316 00:09:05.173492 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"abaf8528-a999-4343-a42b-66283fd65da0","Type":"ContainerStarted","Data":"d1b00e3c05347902f43ed39ea24c3c2c0be4d9421389dcd234e5ed726819f6e2"} Mar 16 00:09:05 crc kubenswrapper[4751]: I0316 00:09:05.191499 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.191478169 podStartE2EDuration="1.191478169s" podCreationTimestamp="2026-03-16 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:09:05.190140475 +0000 UTC m=+173.459472775" watchObservedRunningTime="2026-03-16 00:09:05.191478169 +0000 UTC m=+173.460810459" Mar 16 00:09:06 crc kubenswrapper[4751]: I0316 00:09:06.265738 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c6g5f" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" containerName="registry-server" probeResult="failure" output=< Mar 16 00:09:06 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Mar 16 00:09:06 crc kubenswrapper[4751]: > Mar 16 00:09:06 crc kubenswrapper[4751]: I0316 00:09:06.608443 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 16 00:09:09 crc kubenswrapper[4751]: I0316 00:09:09.199364 4751 generic.go:334] "Generic (PLEG): container finished" podID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerID="91b3e81da9a9f336743af63c56b2ace29ce282f1c0ed420dac4656407b512f51" exitCode=0 Mar 16 00:09:09 crc kubenswrapper[4751]: I0316 00:09:09.199434 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kc6w" event={"ID":"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df","Type":"ContainerDied","Data":"91b3e81da9a9f336743af63c56b2ace29ce282f1c0ed420dac4656407b512f51"} Mar 16 00:09:09 crc kubenswrapper[4751]: I0316 00:09:09.205923 4751 generic.go:334] "Generic (PLEG): container finished" podID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerID="3dd633f9cc74ab824e8193eefb7e5b4b077573c7e246929860b3cf19cdcf7a29" exitCode=0 Mar 16 00:09:09 crc kubenswrapper[4751]: I0316 00:09:09.205985 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc9vc" event={"ID":"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4","Type":"ContainerDied","Data":"3dd633f9cc74ab824e8193eefb7e5b4b077573c7e246929860b3cf19cdcf7a29"} Mar 16 00:09:09 crc kubenswrapper[4751]: I0316 00:09:09.209472 4751 generic.go:334] "Generic (PLEG): container finished" podID="163b6298-4cca-4bea-9272-124555c2b0f4" containerID="5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7" exitCode=0 Mar 16 00:09:09 crc kubenswrapper[4751]: I0316 00:09:09.209549 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chg6k" event={"ID":"163b6298-4cca-4bea-9272-124555c2b0f4","Type":"ContainerDied","Data":"5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7"} Mar 16 00:09:09 crc kubenswrapper[4751]: I0316 00:09:09.213053 4751 generic.go:334] "Generic (PLEG): container finished" podID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerID="eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df" exitCode=0 Mar 16 00:09:09 crc kubenswrapper[4751]: I0316 00:09:09.213121 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbr2k" event={"ID":"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8","Type":"ContainerDied","Data":"eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df"} Mar 16 00:09:09 crc kubenswrapper[4751]: I0316 00:09:09.476328 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 16 00:09:10 crc kubenswrapper[4751]: I0316 00:09:10.496225 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.49620277 podStartE2EDuration="1.49620277s" podCreationTimestamp="2026-03-16 00:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:09:10.492948855 +0000 UTC m=+178.762281185" watchObservedRunningTime="2026-03-16 00:09:10.49620277 +0000 UTC m=+178.765535060" Mar 16 00:09:11 crc kubenswrapper[4751]: I0316 00:09:11.228858 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlrgw" event={"ID":"8f07095b-d178-4ef1-9269-b47e33d71ef3","Type":"ContainerStarted","Data":"6342b3764cb725d4f3ae4e9c388a82ae6908a92cf22be82cb7424a9e7686310b"} Mar 16 00:09:12 crc kubenswrapper[4751]: I0316 00:09:12.241602 4751 generic.go:334] "Generic (PLEG): container finished" podID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerID="6342b3764cb725d4f3ae4e9c388a82ae6908a92cf22be82cb7424a9e7686310b" exitCode=0 Mar 16 00:09:12 crc kubenswrapper[4751]: I0316 00:09:12.241737 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlrgw" event={"ID":"8f07095b-d178-4ef1-9269-b47e33d71ef3","Type":"ContainerDied","Data":"6342b3764cb725d4f3ae4e9c388a82ae6908a92cf22be82cb7424a9e7686310b"} Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.250793 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chg6k" event={"ID":"163b6298-4cca-4bea-9272-124555c2b0f4","Type":"ContainerStarted","Data":"856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671"} Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.254446 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kc6w" event={"ID":"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df","Type":"ContainerStarted","Data":"7c86bd0dab7affbe02238e2fc8a5765d49551ed62988d987b925aa2e3f691f7a"} Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.257088 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc9vc" event={"ID":"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4","Type":"ContainerStarted","Data":"7932415c0e6d1902552b5f52c21327885c0b12df275ce4e58d0a0259b3a42414"} Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.259449 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlrgw" event={"ID":"8f07095b-d178-4ef1-9269-b47e33d71ef3","Type":"ContainerStarted","Data":"30ba8ff4e31de5142bcf76a3e58621e7063f410fb08e9cc5e4c48a96c58c9ac7"} Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.261628 4751 generic.go:334] "Generic (PLEG): container finished" podID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerID="1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa" exitCode=0 Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.261678 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwz46" event={"ID":"bf3ab855-2797-46d2-b42b-e8a43ea455b5","Type":"ContainerDied","Data":"1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa"} Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.266218 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbr2k" event={"ID":"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8","Type":"ContainerStarted","Data":"87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d"} Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.267925 4751 generic.go:334] "Generic (PLEG): container finished" podID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerID="17caa63336ce668a7f6c0d0b5bba1aa4ca99fd48874bb0a1946558ec6e9ab23a" exitCode=0 Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.267952 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqzbn" event={"ID":"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c","Type":"ContainerDied","Data":"17caa63336ce668a7f6c0d0b5bba1aa4ca99fd48874bb0a1946558ec6e9ab23a"} Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.280461 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-chg6k" podStartSLOduration=4.521726559 podStartE2EDuration="52.280451909s" podCreationTimestamp="2026-03-16 00:08:21 +0000 UTC" firstStartedPulling="2026-03-16 00:08:24.382361897 +0000 UTC m=+132.651694187" lastFinishedPulling="2026-03-16 00:09:12.141087207 +0000 UTC m=+180.410419537" observedRunningTime="2026-03-16 00:09:13.276736919 +0000 UTC m=+181.546069209" watchObservedRunningTime="2026-03-16 00:09:13.280451909 +0000 UTC m=+181.549784199" Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.300939 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tbr2k" podStartSLOduration=3.459993116 podStartE2EDuration="52.300915419s" podCreationTimestamp="2026-03-16 00:08:21 +0000 UTC" firstStartedPulling="2026-03-16 00:08:23.344031619 +0000 UTC m=+131.613363909" lastFinishedPulling="2026-03-16 00:09:12.184953882 +0000 UTC m=+180.454286212" observedRunningTime="2026-03-16 00:09:13.296020941 +0000 UTC m=+181.565353231" watchObservedRunningTime="2026-03-16 00:09:13.300915419 +0000 UTC m=+181.570247709" Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.362200 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8kc6w" podStartSLOduration=3.970827708 podStartE2EDuration="52.362172145s" podCreationTimestamp="2026-03-16 00:08:21 +0000 UTC" firstStartedPulling="2026-03-16 00:08:23.273174778 +0000 UTC m=+131.542507058" lastFinishedPulling="2026-03-16 00:09:11.664519205 +0000 UTC m=+179.933851495" observedRunningTime="2026-03-16 00:09:13.339872276 +0000 UTC m=+181.609204566" watchObservedRunningTime="2026-03-16 00:09:13.362172145 +0000 UTC m=+181.631504445" Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.362679 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tlrgw" podStartSLOduration=3.285987888 podStartE2EDuration="49.362672061s" podCreationTimestamp="2026-03-16 00:08:24 +0000 UTC" firstStartedPulling="2026-03-16 00:08:26.618859329 +0000 UTC m=+134.888191619" lastFinishedPulling="2026-03-16 00:09:12.695543502 +0000 UTC m=+180.964875792" observedRunningTime="2026-03-16 00:09:13.359562181 +0000 UTC m=+181.628894471" watchObservedRunningTime="2026-03-16 00:09:13.362672061 +0000 UTC m=+181.632004351" Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.389597 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pc9vc" podStartSLOduration=3.656136391 podStartE2EDuration="50.389578619s" podCreationTimestamp="2026-03-16 00:08:23 +0000 UTC" firstStartedPulling="2026-03-16 00:08:25.536766164 +0000 UTC m=+133.806098454" lastFinishedPulling="2026-03-16 00:09:12.270208382 +0000 UTC m=+180.539540682" observedRunningTime="2026-03-16 00:09:13.388294308 +0000 UTC m=+181.657626598" watchObservedRunningTime="2026-03-16 00:09:13.389578619 +0000 UTC m=+181.658910909" Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.675446 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:09:13 crc kubenswrapper[4751]: I0316 00:09:13.675499 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:09:14 crc kubenswrapper[4751]: I0316 00:09:14.274911 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqzbn" event={"ID":"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c","Type":"ContainerStarted","Data":"f64239b9f399ae12b962432d2a8d18f2c66c15c8e2cbdf468060f94c05b1103b"} Mar 16 00:09:14 crc kubenswrapper[4751]: I0316 00:09:14.277939 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwz46" event={"ID":"bf3ab855-2797-46d2-b42b-e8a43ea455b5","Type":"ContainerStarted","Data":"6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771"} Mar 16 00:09:14 crc kubenswrapper[4751]: I0316 00:09:14.315200 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nwz46" podStartSLOduration=3.158719906 podStartE2EDuration="51.315162595s" podCreationTimestamp="2026-03-16 00:08:23 +0000 UTC" firstStartedPulling="2026-03-16 00:08:25.537140836 +0000 UTC m=+133.806473116" lastFinishedPulling="2026-03-16 00:09:13.693583515 +0000 UTC m=+181.962915805" observedRunningTime="2026-03-16 00:09:14.312014314 +0000 UTC m=+182.581346604" watchObservedRunningTime="2026-03-16 00:09:14.315162595 +0000 UTC m=+182.584494875" Mar 16 00:09:14 crc kubenswrapper[4751]: I0316 00:09:14.315725 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mqzbn" podStartSLOduration=2.985981879 podStartE2EDuration="53.315718893s" podCreationTimestamp="2026-03-16 00:08:21 +0000 UTC" firstStartedPulling="2026-03-16 00:08:23.338348359 +0000 UTC m=+131.607680649" lastFinishedPulling="2026-03-16 00:09:13.668085373 +0000 UTC m=+181.937417663" observedRunningTime="2026-03-16 00:09:14.296371609 +0000 UTC m=+182.565703899" watchObservedRunningTime="2026-03-16 00:09:14.315718893 +0000 UTC m=+182.585051183" Mar 16 00:09:14 crc kubenswrapper[4751]: I0316 00:09:14.650803 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:09:14 crc kubenswrapper[4751]: I0316 00:09:14.650858 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:09:14 crc kubenswrapper[4751]: I0316 00:09:14.736032 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-pc9vc" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerName="registry-server" probeResult="failure" output=< Mar 16 00:09:14 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Mar 16 00:09:14 crc kubenswrapper[4751]: > Mar 16 00:09:15 crc kubenswrapper[4751]: I0316 00:09:15.166176 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:09:15 crc kubenswrapper[4751]: I0316 00:09:15.211011 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:09:15 crc kubenswrapper[4751]: I0316 00:09:15.693232 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tlrgw" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerName="registry-server" probeResult="failure" output=< Mar 16 00:09:15 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Mar 16 00:09:15 crc kubenswrapper[4751]: > Mar 16 00:09:15 crc kubenswrapper[4751]: I0316 00:09:15.894763 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6g5f"] Mar 16 00:09:16 crc kubenswrapper[4751]: I0316 00:09:16.297588 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c6g5f" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" containerName="registry-server" containerID="cri-o://e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24" gracePeriod=2 Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.206042 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.304413 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-utilities\") pod \"c83465bd-08af-460e-ad52-a874fb037bbb\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.304472 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdpxs\" (UniqueName: \"kubernetes.io/projected/c83465bd-08af-460e-ad52-a874fb037bbb-kube-api-access-zdpxs\") pod \"c83465bd-08af-460e-ad52-a874fb037bbb\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.304514 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-catalog-content\") pod \"c83465bd-08af-460e-ad52-a874fb037bbb\" (UID: \"c83465bd-08af-460e-ad52-a874fb037bbb\") " Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.305336 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-utilities" (OuterVolumeSpecName: "utilities") pod "c83465bd-08af-460e-ad52-a874fb037bbb" (UID: "c83465bd-08af-460e-ad52-a874fb037bbb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.311026 4751 generic.go:334] "Generic (PLEG): container finished" podID="c83465bd-08af-460e-ad52-a874fb037bbb" containerID="e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24" exitCode=0 Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.311063 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6g5f" event={"ID":"c83465bd-08af-460e-ad52-a874fb037bbb","Type":"ContainerDied","Data":"e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24"} Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.311141 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6g5f" event={"ID":"c83465bd-08af-460e-ad52-a874fb037bbb","Type":"ContainerDied","Data":"d7cc3cc7da0d2c3f32593a7d1ae675e767d3a81e8f6964eca865e503327d5d4b"} Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.311166 4751 scope.go:117] "RemoveContainer" containerID="e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.311194 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6g5f" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.314847 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c83465bd-08af-460e-ad52-a874fb037bbb-kube-api-access-zdpxs" (OuterVolumeSpecName: "kube-api-access-zdpxs") pod "c83465bd-08af-460e-ad52-a874fb037bbb" (UID: "c83465bd-08af-460e-ad52-a874fb037bbb"). InnerVolumeSpecName "kube-api-access-zdpxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.354773 4751 scope.go:117] "RemoveContainer" containerID="d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.379690 4751 scope.go:117] "RemoveContainer" containerID="8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.392734 4751 scope.go:117] "RemoveContainer" containerID="e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24" Mar 16 00:09:17 crc kubenswrapper[4751]: E0316 00:09:17.393422 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24\": container with ID starting with e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24 not found: ID does not exist" containerID="e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.393461 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24"} err="failed to get container status \"e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24\": rpc error: code = NotFound desc = could not find container \"e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24\": container with ID starting with e8f44e639f0bd1ee4dd96afeebdf1c13922d1d204ace23faed19163c6a178c24 not found: ID does not exist" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.393487 4751 scope.go:117] "RemoveContainer" containerID="d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb" Mar 16 00:09:17 crc kubenswrapper[4751]: E0316 00:09:17.393811 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb\": container with ID starting with d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb not found: ID does not exist" containerID="d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.393880 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb"} err="failed to get container status \"d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb\": rpc error: code = NotFound desc = could not find container \"d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb\": container with ID starting with d6963b8cf404122d9bfe26f03a4ffb612a01e443ff94e01d8a8cf2426cf25dbb not found: ID does not exist" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.393931 4751 scope.go:117] "RemoveContainer" containerID="8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455" Mar 16 00:09:17 crc kubenswrapper[4751]: E0316 00:09:17.394524 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455\": container with ID starting with 8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455 not found: ID does not exist" containerID="8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.394581 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455"} err="failed to get container status \"8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455\": rpc error: code = NotFound desc = could not find container \"8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455\": container with ID starting with 8b3cbb35668e95ec7e46cfd481e7ce160cce829792c5c41a2cbc1ceb0e6f3455 not found: ID does not exist" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.406487 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.406530 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdpxs\" (UniqueName: \"kubernetes.io/projected/c83465bd-08af-460e-ad52-a874fb037bbb-kube-api-access-zdpxs\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.443958 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c83465bd-08af-460e-ad52-a874fb037bbb" (UID: "c83465bd-08af-460e-ad52-a874fb037bbb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.507641 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83465bd-08af-460e-ad52-a874fb037bbb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.648628 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6g5f"] Mar 16 00:09:17 crc kubenswrapper[4751]: I0316 00:09:17.654390 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c6g5f"] Mar 16 00:09:18 crc kubenswrapper[4751]: I0316 00:09:18.463142 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" path="/var/lib/kubelet/pods/c83465bd-08af-460e-ad52-a874fb037bbb/volumes" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.030822 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" podUID="5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" containerName="oauth-openshift" containerID="cri-o://741b041b7d841e9e364dbef29a97a96bc40b6cdb95edc0f376a8cf1cc664ad94" gracePeriod=15 Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.333545 4751 generic.go:334] "Generic (PLEG): container finished" podID="5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" containerID="741b041b7d841e9e364dbef29a97a96bc40b6cdb95edc0f376a8cf1cc664ad94" exitCode=0 Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.333588 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" event={"ID":"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec","Type":"ContainerDied","Data":"741b041b7d841e9e364dbef29a97a96bc40b6cdb95edc0f376a8cf1cc664ad94"} Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.643272 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.752934 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shm2f\" (UniqueName: \"kubernetes.io/projected/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-kube-api-access-shm2f\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.752986 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-cliconfig\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753020 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-error\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753057 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-dir\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753090 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-router-certs\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753136 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-policies\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753156 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-session\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753178 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-trusted-ca-bundle\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753232 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-service-ca\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753254 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-login\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753260 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753305 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-ocp-branding-template\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753410 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-idp-0-file-data\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753457 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-provider-selection\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753514 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-serving-cert\") pod \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\" (UID: \"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec\") " Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.753966 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.754061 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.754083 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.755326 4751 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.755152 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.755568 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.759243 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.759640 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.760091 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.760468 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.761363 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.761457 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-kube-api-access-shm2f" (OuterVolumeSpecName: "kube-api-access-shm2f") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "kube-api-access-shm2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.761637 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.761647 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.773811 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" (UID: "5f11ca1f-26b5-4887-92c5-cdbf4dac90ec"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.856817 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.856882 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.856896 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.856912 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.856925 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.856938 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.856951 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.856963 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shm2f\" (UniqueName: \"kubernetes.io/projected/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-kube-api-access-shm2f\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.856996 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.857009 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.857022 4751 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:20 crc kubenswrapper[4751]: I0316 00:09:20.857036 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.343745 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-9d557989c-t8b7g"] Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.344048 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" podUID="ce637707-a0df-4f7e-83fc-dadbd0857285" containerName="controller-manager" containerID="cri-o://91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e" gracePeriod=30 Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.348017 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" event={"ID":"5f11ca1f-26b5-4887-92c5-cdbf4dac90ec","Type":"ContainerDied","Data":"d7ff99d1f527cf26cc2ce1c8c78a12f0a684da43c048147f7b54109ddd0bc364"} Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.348094 4751 scope.go:117] "RemoveContainer" containerID="741b041b7d841e9e364dbef29a97a96bc40b6cdb95edc0f376a8cf1cc664ad94" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.348291 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sfnlp" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.381739 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp"] Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.381969 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" podUID="58afe499-1e5d-40ea-aa87-78139be5ff44" containerName="route-controller-manager" containerID="cri-o://3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670" gracePeriod=30 Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.420003 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sfnlp"] Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.427904 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sfnlp"] Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.689391 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.689453 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.736949 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.737000 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.764075 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.776733 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.835705 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.835804 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.863750 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.909062 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.974301 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.981894 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-client-ca\") pod \"58afe499-1e5d-40ea-aa87-78139be5ff44\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.981989 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-config\") pod \"58afe499-1e5d-40ea-aa87-78139be5ff44\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.982017 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt6gn\" (UniqueName: \"kubernetes.io/projected/58afe499-1e5d-40ea-aa87-78139be5ff44-kube-api-access-pt6gn\") pod \"58afe499-1e5d-40ea-aa87-78139be5ff44\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.982129 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58afe499-1e5d-40ea-aa87-78139be5ff44-serving-cert\") pod \"58afe499-1e5d-40ea-aa87-78139be5ff44\" (UID: \"58afe499-1e5d-40ea-aa87-78139be5ff44\") " Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.982956 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-client-ca" (OuterVolumeSpecName: "client-ca") pod "58afe499-1e5d-40ea-aa87-78139be5ff44" (UID: "58afe499-1e5d-40ea-aa87-78139be5ff44"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.983267 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-config" (OuterVolumeSpecName: "config") pod "58afe499-1e5d-40ea-aa87-78139be5ff44" (UID: "58afe499-1e5d-40ea-aa87-78139be5ff44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.984027 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.984054 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58afe499-1e5d-40ea-aa87-78139be5ff44-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.986470 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58afe499-1e5d-40ea-aa87-78139be5ff44-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "58afe499-1e5d-40ea-aa87-78139be5ff44" (UID: "58afe499-1e5d-40ea-aa87-78139be5ff44"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:21 crc kubenswrapper[4751]: I0316 00:09:21.987224 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58afe499-1e5d-40ea-aa87-78139be5ff44-kube-api-access-pt6gn" (OuterVolumeSpecName: "kube-api-access-pt6gn") pod "58afe499-1e5d-40ea-aa87-78139be5ff44" (UID: "58afe499-1e5d-40ea-aa87-78139be5ff44"). InnerVolumeSpecName "kube-api-access-pt6gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.084477 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcjf8\" (UniqueName: \"kubernetes.io/projected/ce637707-a0df-4f7e-83fc-dadbd0857285-kube-api-access-tcjf8\") pod \"ce637707-a0df-4f7e-83fc-dadbd0857285\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.084523 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-proxy-ca-bundles\") pod \"ce637707-a0df-4f7e-83fc-dadbd0857285\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.084568 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce637707-a0df-4f7e-83fc-dadbd0857285-serving-cert\") pod \"ce637707-a0df-4f7e-83fc-dadbd0857285\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.084584 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-client-ca\") pod \"ce637707-a0df-4f7e-83fc-dadbd0857285\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.084620 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-config\") pod \"ce637707-a0df-4f7e-83fc-dadbd0857285\" (UID: \"ce637707-a0df-4f7e-83fc-dadbd0857285\") " Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.084831 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt6gn\" (UniqueName: \"kubernetes.io/projected/58afe499-1e5d-40ea-aa87-78139be5ff44-kube-api-access-pt6gn\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.084844 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58afe499-1e5d-40ea-aa87-78139be5ff44-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.085468 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-config" (OuterVolumeSpecName: "config") pod "ce637707-a0df-4f7e-83fc-dadbd0857285" (UID: "ce637707-a0df-4f7e-83fc-dadbd0857285"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.085501 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ce637707-a0df-4f7e-83fc-dadbd0857285" (UID: "ce637707-a0df-4f7e-83fc-dadbd0857285"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.085537 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-client-ca" (OuterVolumeSpecName: "client-ca") pod "ce637707-a0df-4f7e-83fc-dadbd0857285" (UID: "ce637707-a0df-4f7e-83fc-dadbd0857285"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.088590 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce637707-a0df-4f7e-83fc-dadbd0857285-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ce637707-a0df-4f7e-83fc-dadbd0857285" (UID: "ce637707-a0df-4f7e-83fc-dadbd0857285"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.089507 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce637707-a0df-4f7e-83fc-dadbd0857285-kube-api-access-tcjf8" (OuterVolumeSpecName: "kube-api-access-tcjf8") pod "ce637707-a0df-4f7e-83fc-dadbd0857285" (UID: "ce637707-a0df-4f7e-83fc-dadbd0857285"). InnerVolumeSpecName "kube-api-access-tcjf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.144604 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.144666 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.186074 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcjf8\" (UniqueName: \"kubernetes.io/projected/ce637707-a0df-4f7e-83fc-dadbd0857285-kube-api-access-tcjf8\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.186200 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.186221 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce637707-a0df-4f7e-83fc-dadbd0857285-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.186238 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.186255 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce637707-a0df-4f7e-83fc-dadbd0857285-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.215601 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.357916 4751 generic.go:334] "Generic (PLEG): container finished" podID="58afe499-1e5d-40ea-aa87-78139be5ff44" containerID="3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670" exitCode=0 Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.358054 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" event={"ID":"58afe499-1e5d-40ea-aa87-78139be5ff44","Type":"ContainerDied","Data":"3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670"} Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.358085 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.358151 4751 scope.go:117] "RemoveContainer" containerID="3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.358099 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp" event={"ID":"58afe499-1e5d-40ea-aa87-78139be5ff44","Type":"ContainerDied","Data":"563e5a27107ca64ee3b7e277a84770e3c6dc795ac9bb3868192b0fc4a3e80533"} Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.362097 4751 generic.go:334] "Generic (PLEG): container finished" podID="ce637707-a0df-4f7e-83fc-dadbd0857285" containerID="91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e" exitCode=0 Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.362376 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.362415 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" event={"ID":"ce637707-a0df-4f7e-83fc-dadbd0857285","Type":"ContainerDied","Data":"91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e"} Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.363527 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9d557989c-t8b7g" event={"ID":"ce637707-a0df-4f7e-83fc-dadbd0857285","Type":"ContainerDied","Data":"0743de9b9f39b96eb670cac11c4e829c13dfd4e3ecdcc89c9878a2f195931588"} Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.384803 4751 scope.go:117] "RemoveContainer" containerID="3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670" Mar 16 00:09:22 crc kubenswrapper[4751]: E0316 00:09:22.385341 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670\": container with ID starting with 3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670 not found: ID does not exist" containerID="3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.385393 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670"} err="failed to get container status \"3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670\": rpc error: code = NotFound desc = could not find container \"3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670\": container with ID starting with 3ca7e9513bf596030861c8eee5db5e2e4fecace595fd51f0b3e7dad2fd33d670 not found: ID does not exist" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.385423 4751 scope.go:117] "RemoveContainer" containerID="91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.420637 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp"] Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.424508 4751 scope.go:117] "RemoveContainer" containerID="91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e" Mar 16 00:09:22 crc kubenswrapper[4751]: E0316 00:09:22.425179 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e\": container with ID starting with 91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e not found: ID does not exist" containerID="91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.425242 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e"} err="failed to get container status \"91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e\": rpc error: code = NotFound desc = could not find container \"91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e\": container with ID starting with 91f4cab2946e3ad01c3fff397b240d4f74276916db6af2243409a44b38ff2c1e not found: ID does not exist" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.426735 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c874bf775-cg2cp"] Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.433076 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-9d557989c-t8b7g"] Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.433345 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.436611 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.438644 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-9d557989c-t8b7g"] Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.438874 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.442647 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.466643 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58afe499-1e5d-40ea-aa87-78139be5ff44" path="/var/lib/kubelet/pods/58afe499-1e5d-40ea-aa87-78139be5ff44/volumes" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.467598 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" path="/var/lib/kubelet/pods/5f11ca1f-26b5-4887-92c5-cdbf4dac90ec/volumes" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.469061 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce637707-a0df-4f7e-83fc-dadbd0857285" path="/var/lib/kubelet/pods/ce637707-a0df-4f7e-83fc-dadbd0857285/volumes" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521341 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn"] Mar 16 00:09:22 crc kubenswrapper[4751]: E0316 00:09:22.521617 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" containerName="oauth-openshift" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521633 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" containerName="oauth-openshift" Mar 16 00:09:22 crc kubenswrapper[4751]: E0316 00:09:22.521648 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" containerName="extract-content" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521656 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" containerName="extract-content" Mar 16 00:09:22 crc kubenswrapper[4751]: E0316 00:09:22.521670 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" containerName="registry-server" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521677 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" containerName="registry-server" Mar 16 00:09:22 crc kubenswrapper[4751]: E0316 00:09:22.521694 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" containerName="extract-utilities" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521703 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" containerName="extract-utilities" Mar 16 00:09:22 crc kubenswrapper[4751]: E0316 00:09:22.521713 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce637707-a0df-4f7e-83fc-dadbd0857285" containerName="controller-manager" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521722 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce637707-a0df-4f7e-83fc-dadbd0857285" containerName="controller-manager" Mar 16 00:09:22 crc kubenswrapper[4751]: E0316 00:09:22.521736 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58afe499-1e5d-40ea-aa87-78139be5ff44" containerName="route-controller-manager" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521743 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="58afe499-1e5d-40ea-aa87-78139be5ff44" containerName="route-controller-manager" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521893 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="58afe499-1e5d-40ea-aa87-78139be5ff44" containerName="route-controller-manager" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521905 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f11ca1f-26b5-4887-92c5-cdbf4dac90ec" containerName="oauth-openshift" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521917 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="c83465bd-08af-460e-ad52-a874fb037bbb" containerName="registry-server" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.521931 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce637707-a0df-4f7e-83fc-dadbd0857285" containerName="controller-manager" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.522892 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.525951 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-d669cf5cf-5s5gd"] Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.526139 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.526492 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.526670 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.526846 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.526903 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.527226 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.527586 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.532830 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.533093 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.534018 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.534209 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.534812 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.535045 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.537815 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn"] Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.540916 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d669cf5cf-5s5gd"] Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.546448 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.590224 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/642903d1-a399-42c9-94f3-97eb5c4ca96e-serving-cert\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.590267 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfdqk\" (UniqueName: \"kubernetes.io/projected/b469eae7-c974-43b0-bf74-706cc2bf62aa-kube-api-access-tfdqk\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.590321 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-client-ca\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.590429 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-proxy-ca-bundles\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.590460 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-client-ca\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.590489 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b469eae7-c974-43b0-bf74-706cc2bf62aa-serving-cert\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.590512 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-config\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.590696 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95wbw\" (UniqueName: \"kubernetes.io/projected/642903d1-a399-42c9-94f3-97eb5c4ca96e-kube-api-access-95wbw\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.590758 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-config\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.691969 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95wbw\" (UniqueName: \"kubernetes.io/projected/642903d1-a399-42c9-94f3-97eb5c4ca96e-kube-api-access-95wbw\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.692042 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-config\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.692094 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/642903d1-a399-42c9-94f3-97eb5c4ca96e-serving-cert\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.692165 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfdqk\" (UniqueName: \"kubernetes.io/projected/b469eae7-c974-43b0-bf74-706cc2bf62aa-kube-api-access-tfdqk\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.692242 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-client-ca\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.692300 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-proxy-ca-bundles\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.692355 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-client-ca\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.692415 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b469eae7-c974-43b0-bf74-706cc2bf62aa-serving-cert\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.692465 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-config\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.694042 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-client-ca\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.694358 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-proxy-ca-bundles\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.694419 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-client-ca\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.695658 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-config\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.695748 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-config\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.699507 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b469eae7-c974-43b0-bf74-706cc2bf62aa-serving-cert\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.705446 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/642903d1-a399-42c9-94f3-97eb5c4ca96e-serving-cert\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.723737 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfdqk\" (UniqueName: \"kubernetes.io/projected/b469eae7-c974-43b0-bf74-706cc2bf62aa-kube-api-access-tfdqk\") pod \"controller-manager-d669cf5cf-5s5gd\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.724858 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95wbw\" (UniqueName: \"kubernetes.io/projected/642903d1-a399-42c9-94f3-97eb5c4ca96e-kube-api-access-95wbw\") pod \"route-controller-manager-598cd6b6f-dbnwn\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.853062 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:22 crc kubenswrapper[4751]: I0316 00:09:22.865238 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.174431 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d669cf5cf-5s5gd"] Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.239535 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn"] Mar 16 00:09:23 crc kubenswrapper[4751]: W0316 00:09:23.243766 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod642903d1_a399_42c9_94f3_97eb5c4ca96e.slice/crio-9082f5058320c4abc3ae15d5557d8c070f9e5b8d5521afc7d4cddf396334a923 WatchSource:0}: Error finding container 9082f5058320c4abc3ae15d5557d8c070f9e5b8d5521afc7d4cddf396334a923: Status 404 returned error can't find the container with id 9082f5058320c4abc3ae15d5557d8c070f9e5b8d5521afc7d4cddf396334a923 Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.372202 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" event={"ID":"642903d1-a399-42c9-94f3-97eb5c4ca96e","Type":"ContainerStarted","Data":"6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d"} Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.373247 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" event={"ID":"642903d1-a399-42c9-94f3-97eb5c4ca96e","Type":"ContainerStarted","Data":"9082f5058320c4abc3ae15d5557d8c070f9e5b8d5521afc7d4cddf396334a923"} Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.373317 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.373617 4751 patch_prober.go:28] interesting pod/route-controller-manager-598cd6b6f-dbnwn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" start-of-body= Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.373658 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.373666 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" event={"ID":"b469eae7-c974-43b0-bf74-706cc2bf62aa","Type":"ContainerStarted","Data":"4bf2b3eaacd5da6e33cce69b20f9bcca9fd90a306f9cf85979ef3d1d533258cd"} Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.373727 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" event={"ID":"b469eae7-c974-43b0-bf74-706cc2bf62aa","Type":"ContainerStarted","Data":"cb3dd34b06bd04a8adad3eedc01959f2f03d6e631a3c66ce18245295578ca726"} Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.374154 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.377238 4751 patch_prober.go:28] interesting pod/controller-manager-d669cf5cf-5s5gd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.377291 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.395863 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" podStartSLOduration=2.3958430330000002 podStartE2EDuration="2.395843033s" podCreationTimestamp="2026-03-16 00:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:09:23.390364226 +0000 UTC m=+191.659696536" watchObservedRunningTime="2026-03-16 00:09:23.395843033 +0000 UTC m=+191.665175323" Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.409949 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" podStartSLOduration=2.4099325670000002 podStartE2EDuration="2.409932567s" podCreationTimestamp="2026-03-16 00:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:09:23.408886064 +0000 UTC m=+191.678218354" watchObservedRunningTime="2026-03-16 00:09:23.409932567 +0000 UTC m=+191.679264877" Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.718979 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.759611 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:09:23 crc kubenswrapper[4751]: I0316 00:09:23.889986 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-chg6k"] Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.047247 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.047302 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.091309 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tbr2k"] Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.120620 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.385797 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-chg6k" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" containerName="registry-server" containerID="cri-o://856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671" gracePeriod=2 Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.387097 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tbr2k" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerName="registry-server" containerID="cri-o://87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d" gracePeriod=2 Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.398013 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.398724 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.477931 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.698033 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.753541 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.830733 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.885708 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.920400 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d87fq\" (UniqueName: \"kubernetes.io/projected/163b6298-4cca-4bea-9272-124555c2b0f4-kube-api-access-d87fq\") pod \"163b6298-4cca-4bea-9272-124555c2b0f4\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.920478 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlr45\" (UniqueName: \"kubernetes.io/projected/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-kube-api-access-rlr45\") pod \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.920506 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-catalog-content\") pod \"163b6298-4cca-4bea-9272-124555c2b0f4\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.921381 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-utilities\") pod \"163b6298-4cca-4bea-9272-124555c2b0f4\" (UID: \"163b6298-4cca-4bea-9272-124555c2b0f4\") " Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.921517 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-utilities\") pod \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.921627 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-catalog-content\") pod \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\" (UID: \"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8\") " Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.922522 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-utilities" (OuterVolumeSpecName: "utilities") pod "f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" (UID: "f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.922942 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-utilities" (OuterVolumeSpecName: "utilities") pod "163b6298-4cca-4bea-9272-124555c2b0f4" (UID: "163b6298-4cca-4bea-9272-124555c2b0f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.926701 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-kube-api-access-rlr45" (OuterVolumeSpecName: "kube-api-access-rlr45") pod "f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" (UID: "f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8"). InnerVolumeSpecName "kube-api-access-rlr45". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.927384 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/163b6298-4cca-4bea-9272-124555c2b0f4-kube-api-access-d87fq" (OuterVolumeSpecName: "kube-api-access-d87fq") pod "163b6298-4cca-4bea-9272-124555c2b0f4" (UID: "163b6298-4cca-4bea-9272-124555c2b0f4"). InnerVolumeSpecName "kube-api-access-d87fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.978477 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "163b6298-4cca-4bea-9272-124555c2b0f4" (UID: "163b6298-4cca-4bea-9272-124555c2b0f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:09:24 crc kubenswrapper[4751]: I0316 00:09:24.990451 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" (UID: "f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.023331 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.023355 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.023365 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.023379 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d87fq\" (UniqueName: \"kubernetes.io/projected/163b6298-4cca-4bea-9272-124555c2b0f4-kube-api-access-d87fq\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.023390 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlr45\" (UniqueName: \"kubernetes.io/projected/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8-kube-api-access-rlr45\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.023398 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/163b6298-4cca-4bea-9272-124555c2b0f4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.399379 4751 generic.go:334] "Generic (PLEG): container finished" podID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerID="87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d" exitCode=0 Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.399506 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbr2k" event={"ID":"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8","Type":"ContainerDied","Data":"87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d"} Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.399549 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbr2k" event={"ID":"f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8","Type":"ContainerDied","Data":"fd718e5457525a1d4645330b108810565fed74d8324c17ea0aba899211525a35"} Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.399580 4751 scope.go:117] "RemoveContainer" containerID="87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.399764 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tbr2k" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.406120 4751 generic.go:334] "Generic (PLEG): container finished" podID="163b6298-4cca-4bea-9272-124555c2b0f4" containerID="856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671" exitCode=0 Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.406576 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chg6k" event={"ID":"163b6298-4cca-4bea-9272-124555c2b0f4","Type":"ContainerDied","Data":"856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671"} Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.406654 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-chg6k" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.406700 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chg6k" event={"ID":"163b6298-4cca-4bea-9272-124555c2b0f4","Type":"ContainerDied","Data":"2b723c23a50e62d87c006b204cf4704ab5176d57f415f4c92e7e6618ff41685d"} Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.421086 4751 scope.go:117] "RemoveContainer" containerID="eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.463564 4751 scope.go:117] "RemoveContainer" containerID="4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.467044 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-chg6k"] Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.473864 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-chg6k"] Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.487868 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tbr2k"] Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.491756 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tbr2k"] Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.549499 4751 scope.go:117] "RemoveContainer" containerID="87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d" Mar 16 00:09:25 crc kubenswrapper[4751]: E0316 00:09:25.551803 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d\": container with ID starting with 87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d not found: ID does not exist" containerID="87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.551847 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d"} err="failed to get container status \"87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d\": rpc error: code = NotFound desc = could not find container \"87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d\": container with ID starting with 87ece864a47e521b10220c4ff4bd404cc9982a9ac00a1679392cc8c39f9f0e2d not found: ID does not exist" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.551882 4751 scope.go:117] "RemoveContainer" containerID="eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df" Mar 16 00:09:25 crc kubenswrapper[4751]: E0316 00:09:25.552473 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df\": container with ID starting with eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df not found: ID does not exist" containerID="eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.552555 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df"} err="failed to get container status \"eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df\": rpc error: code = NotFound desc = could not find container \"eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df\": container with ID starting with eca0084557522b8eee417a30113f6fa4a52e3e79f455b68278dd194b220d40df not found: ID does not exist" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.552614 4751 scope.go:117] "RemoveContainer" containerID="4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa" Mar 16 00:09:25 crc kubenswrapper[4751]: E0316 00:09:25.553172 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa\": container with ID starting with 4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa not found: ID does not exist" containerID="4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.553207 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa"} err="failed to get container status \"4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa\": rpc error: code = NotFound desc = could not find container \"4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa\": container with ID starting with 4852c23e4010a81fc1b4fb29545a3fd11285404caaf234e5e1086e69cc6762aa not found: ID does not exist" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.553227 4751 scope.go:117] "RemoveContainer" containerID="856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.586300 4751 scope.go:117] "RemoveContainer" containerID="5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.606836 4751 scope.go:117] "RemoveContainer" containerID="345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.637973 4751 scope.go:117] "RemoveContainer" containerID="856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671" Mar 16 00:09:25 crc kubenswrapper[4751]: E0316 00:09:25.639232 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671\": container with ID starting with 856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671 not found: ID does not exist" containerID="856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.639317 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671"} err="failed to get container status \"856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671\": rpc error: code = NotFound desc = could not find container \"856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671\": container with ID starting with 856cdefb10fc38eb1f712208998945f1ca13779ab74c906be751cbad8a61d671 not found: ID does not exist" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.639366 4751 scope.go:117] "RemoveContainer" containerID="5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7" Mar 16 00:09:25 crc kubenswrapper[4751]: E0316 00:09:25.639939 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7\": container with ID starting with 5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7 not found: ID does not exist" containerID="5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.639981 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7"} err="failed to get container status \"5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7\": rpc error: code = NotFound desc = could not find container \"5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7\": container with ID starting with 5ad1337b91361a6b91c8d068c088cc293712b1164ead7670c7212b5cb6d548d7 not found: ID does not exist" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.640013 4751 scope.go:117] "RemoveContainer" containerID="345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3" Mar 16 00:09:25 crc kubenswrapper[4751]: E0316 00:09:25.640567 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3\": container with ID starting with 345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3 not found: ID does not exist" containerID="345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3" Mar 16 00:09:25 crc kubenswrapper[4751]: I0316 00:09:25.640604 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3"} err="failed to get container status \"345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3\": rpc error: code = NotFound desc = could not find container \"345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3\": container with ID starting with 345d19d6bdbfe7491a17834d288d9740f665c56c93c7377b450fadaa5447d1c3 not found: ID does not exist" Mar 16 00:09:26 crc kubenswrapper[4751]: I0316 00:09:26.293510 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwz46"] Mar 16 00:09:26 crc kubenswrapper[4751]: I0316 00:09:26.417413 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nwz46" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerName="registry-server" containerID="cri-o://6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771" gracePeriod=2 Mar 16 00:09:26 crc kubenswrapper[4751]: I0316 00:09:26.473412 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" path="/var/lib/kubelet/pods/163b6298-4cca-4bea-9272-124555c2b0f4/volumes" Mar 16 00:09:26 crc kubenswrapper[4751]: I0316 00:09:26.474346 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" path="/var/lib/kubelet/pods/f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8/volumes" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.009749 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.156920 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-utilities\") pod \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.157062 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxght\" (UniqueName: \"kubernetes.io/projected/bf3ab855-2797-46d2-b42b-e8a43ea455b5-kube-api-access-mxght\") pod \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.157286 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-catalog-content\") pod \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\" (UID: \"bf3ab855-2797-46d2-b42b-e8a43ea455b5\") " Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.158682 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-utilities" (OuterVolumeSpecName: "utilities") pod "bf3ab855-2797-46d2-b42b-e8a43ea455b5" (UID: "bf3ab855-2797-46d2-b42b-e8a43ea455b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.162230 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf3ab855-2797-46d2-b42b-e8a43ea455b5-kube-api-access-mxght" (OuterVolumeSpecName: "kube-api-access-mxght") pod "bf3ab855-2797-46d2-b42b-e8a43ea455b5" (UID: "bf3ab855-2797-46d2-b42b-e8a43ea455b5"). InnerVolumeSpecName "kube-api-access-mxght". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.206757 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf3ab855-2797-46d2-b42b-e8a43ea455b5" (UID: "bf3ab855-2797-46d2-b42b-e8a43ea455b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.259707 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxght\" (UniqueName: \"kubernetes.io/projected/bf3ab855-2797-46d2-b42b-e8a43ea455b5-kube-api-access-mxght\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.259755 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.259773 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf3ab855-2797-46d2-b42b-e8a43ea455b5-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.426532 4751 generic.go:334] "Generic (PLEG): container finished" podID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerID="6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771" exitCode=0 Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.426600 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwz46" event={"ID":"bf3ab855-2797-46d2-b42b-e8a43ea455b5","Type":"ContainerDied","Data":"6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771"} Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.426673 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwz46" event={"ID":"bf3ab855-2797-46d2-b42b-e8a43ea455b5","Type":"ContainerDied","Data":"3d6e1b14cf44a97e461b30584f6618c2f46d4bf5b49e1c8ee3932eb7c845a7f3"} Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.426623 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwz46" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.426717 4751 scope.go:117] "RemoveContainer" containerID="6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.459336 4751 scope.go:117] "RemoveContainer" containerID="1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.485836 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwz46"] Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.494609 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwz46"] Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.498830 4751 scope.go:117] "RemoveContainer" containerID="2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.529531 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-f77784dbc-7mjms"] Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.530385 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" containerName="extract-utilities" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.531144 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" containerName="extract-utilities" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.531324 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerName="registry-server" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.531455 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerName="registry-server" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.531571 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerName="registry-server" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.531678 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerName="registry-server" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.531792 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerName="extract-utilities" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.531912 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerName="extract-utilities" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.532043 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" containerName="registry-server" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.532188 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" containerName="registry-server" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.532334 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerName="extract-utilities" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.532570 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerName="extract-utilities" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.532720 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerName="extract-content" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.532871 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerName="extract-content" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.533017 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" containerName="extract-content" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.533232 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" containerName="extract-content" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.533467 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerName="extract-content" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.533602 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerName="extract-content" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.533969 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" containerName="registry-server" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.534195 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f045bf0b-fcd6-44a3-a21a-65c6c5e54aa8" containerName="registry-server" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.534412 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="163b6298-4cca-4bea-9272-124555c2b0f4" containerName="registry-server" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.535176 4751 scope.go:117] "RemoveContainer" containerID="6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.535494 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.538211 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771\": container with ID starting with 6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771 not found: ID does not exist" containerID="6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.538319 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771"} err="failed to get container status \"6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771\": rpc error: code = NotFound desc = could not find container \"6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771\": container with ID starting with 6888c944af9c6b0bd77af5ef5b105f47a36ba55ad7bf5247939d12a94db1b771 not found: ID does not exist" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.538360 4751 scope.go:117] "RemoveContainer" containerID="1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.542787 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.552343 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa\": container with ID starting with 1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa not found: ID does not exist" containerID="1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.552416 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa"} err="failed to get container status \"1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa\": rpc error: code = NotFound desc = could not find container \"1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa\": container with ID starting with 1e8238ee62e7915988c231840b8349fef830fe30e5c3f0e79e049cb1e7db78fa not found: ID does not exist" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.552460 4751 scope.go:117] "RemoveContainer" containerID="2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6" Mar 16 00:09:27 crc kubenswrapper[4751]: E0316 00:09:27.553184 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6\": container with ID starting with 2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6 not found: ID does not exist" containerID="2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.553370 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.553545 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6"} err="failed to get container status \"2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6\": rpc error: code = NotFound desc = could not find container \"2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6\": container with ID starting with 2e793c68a10ca5e289d232945ecfeaa50e050db523bdcca6c1e81f9150e46fd6 not found: ID does not exist" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.557544 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.557935 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.558145 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.558165 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.558231 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.557947 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.558024 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.558042 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.558545 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.559037 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.563396 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.563970 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-f77784dbc-7mjms"] Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.565979 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-template-login\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.566264 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.566451 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-template-error\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.566686 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-router-certs\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.566835 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-service-ca\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.567054 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.567246 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-serving-cert\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.567439 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.567626 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-audit-policies\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.567819 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-cliconfig\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.567992 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a3b1197-2866-4144-8848-bb76db205a64-audit-dir\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.568176 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.568425 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-session\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.568626 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k9kq\" (UniqueName: \"kubernetes.io/projected/3a3b1197-2866-4144-8848-bb76db205a64-kube-api-access-2k9kq\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.572951 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.589546 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.670493 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-cliconfig\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.670856 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a3b1197-2866-4144-8848-bb76db205a64-audit-dir\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.671057 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a3b1197-2866-4144-8848-bb76db205a64-audit-dir\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.671304 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.671523 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-session\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.671702 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-cliconfig\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.671888 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k9kq\" (UniqueName: \"kubernetes.io/projected/3a3b1197-2866-4144-8848-bb76db205a64-kube-api-access-2k9kq\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.676311 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.677758 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-session\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.678007 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-template-login\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.679335 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-template-login\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.678438 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.680375 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-template-error\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.680987 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-service-ca\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.681516 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-router-certs\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.681738 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-serving-cert\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.681916 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.682091 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.682328 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-audit-policies\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.683896 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-audit-policies\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.684089 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-template-error\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.684356 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-service-ca\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.685019 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.687157 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.687656 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-serving-cert\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.687892 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.688148 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3a3b1197-2866-4144-8848-bb76db205a64-v4-0-config-system-router-certs\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.700228 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k9kq\" (UniqueName: \"kubernetes.io/projected/3a3b1197-2866-4144-8848-bb76db205a64-kube-api-access-2k9kq\") pod \"oauth-openshift-f77784dbc-7mjms\" (UID: \"3a3b1197-2866-4144-8848-bb76db205a64\") " pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:27 crc kubenswrapper[4751]: I0316 00:09:27.865244 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:28 crc kubenswrapper[4751]: I0316 00:09:28.294273 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-f77784dbc-7mjms"] Mar 16 00:09:28 crc kubenswrapper[4751]: I0316 00:09:28.438495 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" event={"ID":"3a3b1197-2866-4144-8848-bb76db205a64","Type":"ContainerStarted","Data":"6a628feaacea5e016f36a1e1495701e0fb6a2e66fcf4208737f6fafce7fe7b6a"} Mar 16 00:09:28 crc kubenswrapper[4751]: I0316 00:09:28.463468 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf3ab855-2797-46d2-b42b-e8a43ea455b5" path="/var/lib/kubelet/pods/bf3ab855-2797-46d2-b42b-e8a43ea455b5/volumes" Mar 16 00:09:29 crc kubenswrapper[4751]: I0316 00:09:29.447712 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" event={"ID":"3a3b1197-2866-4144-8848-bb76db205a64","Type":"ContainerStarted","Data":"4e5b58c37a0f20c41ab403709b13442615271b90805f08e461557e3d611e641f"} Mar 16 00:09:29 crc kubenswrapper[4751]: I0316 00:09:29.449989 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:29 crc kubenswrapper[4751]: I0316 00:09:29.461829 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" Mar 16 00:09:29 crc kubenswrapper[4751]: I0316 00:09:29.477845 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-f77784dbc-7mjms" podStartSLOduration=34.477811994 podStartE2EDuration="34.477811994s" podCreationTimestamp="2026-03-16 00:08:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:09:29.475726957 +0000 UTC m=+197.745059257" watchObservedRunningTime="2026-03-16 00:09:29.477811994 +0000 UTC m=+197.747144294" Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.353316 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-d669cf5cf-5s5gd"] Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.354422 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" containerName="controller-manager" containerID="cri-o://4bf2b3eaacd5da6e33cce69b20f9bcca9fd90a306f9cf85979ef3d1d533258cd" gracePeriod=30 Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.439873 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn"] Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.440431 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" containerName="route-controller-manager" containerID="cri-o://6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d" gracePeriod=30 Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.523634 4751 generic.go:334] "Generic (PLEG): container finished" podID="b469eae7-c974-43b0-bf74-706cc2bf62aa" containerID="4bf2b3eaacd5da6e33cce69b20f9bcca9fd90a306f9cf85979ef3d1d533258cd" exitCode=0 Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.523679 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" event={"ID":"b469eae7-c974-43b0-bf74-706cc2bf62aa","Type":"ContainerDied","Data":"4bf2b3eaacd5da6e33cce69b20f9bcca9fd90a306f9cf85979ef3d1d533258cd"} Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.935795 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.987896 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/642903d1-a399-42c9-94f3-97eb5c4ca96e-serving-cert\") pod \"642903d1-a399-42c9-94f3-97eb5c4ca96e\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.988372 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95wbw\" (UniqueName: \"kubernetes.io/projected/642903d1-a399-42c9-94f3-97eb5c4ca96e-kube-api-access-95wbw\") pod \"642903d1-a399-42c9-94f3-97eb5c4ca96e\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.988449 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-config\") pod \"642903d1-a399-42c9-94f3-97eb5c4ca96e\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.988547 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-client-ca\") pod \"642903d1-a399-42c9-94f3-97eb5c4ca96e\" (UID: \"642903d1-a399-42c9-94f3-97eb5c4ca96e\") " Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.989703 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-client-ca" (OuterVolumeSpecName: "client-ca") pod "642903d1-a399-42c9-94f3-97eb5c4ca96e" (UID: "642903d1-a399-42c9-94f3-97eb5c4ca96e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.993659 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-config" (OuterVolumeSpecName: "config") pod "642903d1-a399-42c9-94f3-97eb5c4ca96e" (UID: "642903d1-a399-42c9-94f3-97eb5c4ca96e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.997702 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642903d1-a399-42c9-94f3-97eb5c4ca96e-kube-api-access-95wbw" (OuterVolumeSpecName: "kube-api-access-95wbw") pod "642903d1-a399-42c9-94f3-97eb5c4ca96e" (UID: "642903d1-a399-42c9-94f3-97eb5c4ca96e"). InnerVolumeSpecName "kube-api-access-95wbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:41 crc kubenswrapper[4751]: I0316 00:09:41.997945 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642903d1-a399-42c9-94f3-97eb5c4ca96e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "642903d1-a399-42c9-94f3-97eb5c4ca96e" (UID: "642903d1-a399-42c9-94f3-97eb5c4ca96e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.002708 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.089726 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfdqk\" (UniqueName: \"kubernetes.io/projected/b469eae7-c974-43b0-bf74-706cc2bf62aa-kube-api-access-tfdqk\") pod \"b469eae7-c974-43b0-bf74-706cc2bf62aa\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.089848 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-config\") pod \"b469eae7-c974-43b0-bf74-706cc2bf62aa\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.089880 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-client-ca\") pod \"b469eae7-c974-43b0-bf74-706cc2bf62aa\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.089926 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-proxy-ca-bundles\") pod \"b469eae7-c974-43b0-bf74-706cc2bf62aa\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.089974 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b469eae7-c974-43b0-bf74-706cc2bf62aa-serving-cert\") pod \"b469eae7-c974-43b0-bf74-706cc2bf62aa\" (UID: \"b469eae7-c974-43b0-bf74-706cc2bf62aa\") " Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.090308 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.090334 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/642903d1-a399-42c9-94f3-97eb5c4ca96e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.090346 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95wbw\" (UniqueName: \"kubernetes.io/projected/642903d1-a399-42c9-94f3-97eb5c4ca96e-kube-api-access-95wbw\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.090359 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/642903d1-a399-42c9-94f3-97eb5c4ca96e-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.091600 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-client-ca" (OuterVolumeSpecName: "client-ca") pod "b469eae7-c974-43b0-bf74-706cc2bf62aa" (UID: "b469eae7-c974-43b0-bf74-706cc2bf62aa"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.091622 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-config" (OuterVolumeSpecName: "config") pod "b469eae7-c974-43b0-bf74-706cc2bf62aa" (UID: "b469eae7-c974-43b0-bf74-706cc2bf62aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.091616 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b469eae7-c974-43b0-bf74-706cc2bf62aa" (UID: "b469eae7-c974-43b0-bf74-706cc2bf62aa"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.093613 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b469eae7-c974-43b0-bf74-706cc2bf62aa-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b469eae7-c974-43b0-bf74-706cc2bf62aa" (UID: "b469eae7-c974-43b0-bf74-706cc2bf62aa"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.094197 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b469eae7-c974-43b0-bf74-706cc2bf62aa-kube-api-access-tfdqk" (OuterVolumeSpecName: "kube-api-access-tfdqk") pod "b469eae7-c974-43b0-bf74-706cc2bf62aa" (UID: "b469eae7-c974-43b0-bf74-706cc2bf62aa"). InnerVolumeSpecName "kube-api-access-tfdqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.192110 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.192168 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-client-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.192184 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b469eae7-c974-43b0-bf74-706cc2bf62aa-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.192198 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b469eae7-c974-43b0-bf74-706cc2bf62aa-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.192212 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfdqk\" (UniqueName: \"kubernetes.io/projected/b469eae7-c974-43b0-bf74-706cc2bf62aa-kube-api-access-tfdqk\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.422250 4751 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.422616 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a" gracePeriod=15 Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.422787 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807" gracePeriod=15 Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.422835 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a" gracePeriod=15 Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.422871 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8" gracePeriod=15 Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.422909 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d" gracePeriod=15 Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.424722 4751 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425095 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425147 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425169 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425183 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425200 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" containerName="controller-manager" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425213 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" containerName="controller-manager" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425231 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425243 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425258 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425270 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425285 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425298 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425311 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425323 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425336 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425350 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425373 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" containerName="route-controller-manager" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425390 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" containerName="route-controller-manager" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425423 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425436 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425450 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425462 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425620 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425640 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" containerName="controller-manager" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425657 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425669 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" containerName="route-controller-manager" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425695 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425713 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425727 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425741 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425754 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.425948 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.425964 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.426168 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.426196 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.428000 4751 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.428970 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.434531 4751 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.495747 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.495823 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.495854 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.495884 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.495955 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.495971 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.495986 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.496004 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.508930 4751 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.36:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.535573 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" event={"ID":"b469eae7-c974-43b0-bf74-706cc2bf62aa","Type":"ContainerDied","Data":"cb3dd34b06bd04a8adad3eedc01959f2f03d6e631a3c66ce18245295578ca726"} Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.535627 4751 scope.go:117] "RemoveContainer" containerID="4bf2b3eaacd5da6e33cce69b20f9bcca9fd90a306f9cf85979ef3d1d533258cd" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.535747 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.536878 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.540335 4751 generic.go:334] "Generic (PLEG): container finished" podID="642903d1-a399-42c9-94f3-97eb5c4ca96e" containerID="6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d" exitCode=0 Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.540382 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" event={"ID":"642903d1-a399-42c9-94f3-97eb5c4ca96e","Type":"ContainerDied","Data":"6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d"} Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.540412 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" event={"ID":"642903d1-a399-42c9-94f3-97eb5c4ca96e","Type":"ContainerDied","Data":"9082f5058320c4abc3ae15d5557d8c070f9e5b8d5521afc7d4cddf396334a923"} Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.540489 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.541979 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.542555 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.578832 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.579197 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.579457 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.579647 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.590622 4751 scope.go:117] "RemoveContainer" containerID="6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597270 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597312 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597335 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597354 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597360 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597382 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597398 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597440 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597484 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597516 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597523 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597553 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597424 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597600 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597569 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.597686 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.611583 4751 scope.go:117] "RemoveContainer" containerID="6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d" Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.611930 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d\": container with ID starting with 6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d not found: ID does not exist" containerID="6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.611961 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d"} err="failed to get container status \"6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d\": rpc error: code = NotFound desc = could not find container \"6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d\": container with ID starting with 6b46f926f756898f1aac85653b9b595e2ac8e82ccf8f5114315208e22651722d not found: ID does not exist" Mar 16 00:09:42 crc kubenswrapper[4751]: I0316 00:09:42.810241 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:42 crc kubenswrapper[4751]: W0316 00:09:42.831697 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-4992b9fd5d2b798a7d0a6517b4b94463b8f134b18e2d9c278c8049339ef69fbe WatchSource:0}: Error finding container 4992b9fd5d2b798a7d0a6517b4b94463b8f134b18e2d9c278c8049339ef69fbe: Status 404 returned error can't find the container with id 4992b9fd5d2b798a7d0a6517b4b94463b8f134b18e2d9c278c8049339ef69fbe Mar 16 00:09:42 crc kubenswrapper[4751]: E0316 00:09:42.835424 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.36:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189d29cba279d130 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:09:42.83483576 +0000 UTC m=+211.104168070,LastTimestamp:2026-03-16 00:09:42.83483576 +0000 UTC m=+211.104168070,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.556538 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.558982 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.560161 4751 scope.go:117] "RemoveContainer" containerID="b69aec7ebf2673e18b14af4ef8f86ad55e123418dde4f0eaf8d136d41610f9dd" Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.560093 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807" exitCode=0 Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.560306 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a" exitCode=0 Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.560330 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8" exitCode=0 Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.560341 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d" exitCode=2 Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.569387 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"3066c873ba83a58e27cc3b8b9c8be7a3ab903a30be5e955869b15921476ef076"} Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.569488 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"4992b9fd5d2b798a7d0a6517b4b94463b8f134b18e2d9c278c8049339ef69fbe"} Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.570579 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:43 crc kubenswrapper[4751]: E0316 00:09:43.570631 4751 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.36:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.571261 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.571984 4751 generic.go:334] "Generic (PLEG): container finished" podID="abaf8528-a999-4343-a42b-66283fd65da0" containerID="f598c75bdb47a637122ffd75060f5d76ca629abf1c71116d823ba544c7dff990" exitCode=0 Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.572056 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"abaf8528-a999-4343-a42b-66283fd65da0","Type":"ContainerDied","Data":"f598c75bdb47a637122ffd75060f5d76ca629abf1c71116d823ba544c7dff990"} Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.573073 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.573722 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:43 crc kubenswrapper[4751]: I0316 00:09:43.574431 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.590005 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.785505 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.786659 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.787363 4751 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.787859 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.788180 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.788549 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.844927 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.844991 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.845007 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.845068 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.845083 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.845192 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.845376 4751 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.845402 4751 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.845414 4751 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.862630 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.863283 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.863727 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.864134 4751 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.864480 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.946037 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-kubelet-dir\") pod \"abaf8528-a999-4343-a42b-66283fd65da0\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.946155 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abaf8528-a999-4343-a42b-66283fd65da0-kube-api-access\") pod \"abaf8528-a999-4343-a42b-66283fd65da0\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.946211 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-var-lock\") pod \"abaf8528-a999-4343-a42b-66283fd65da0\" (UID: \"abaf8528-a999-4343-a42b-66283fd65da0\") " Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.946304 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "abaf8528-a999-4343-a42b-66283fd65da0" (UID: "abaf8528-a999-4343-a42b-66283fd65da0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.946382 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-var-lock" (OuterVolumeSpecName: "var-lock") pod "abaf8528-a999-4343-a42b-66283fd65da0" (UID: "abaf8528-a999-4343-a42b-66283fd65da0"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:09:44 crc kubenswrapper[4751]: I0316 00:09:44.954879 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abaf8528-a999-4343-a42b-66283fd65da0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "abaf8528-a999-4343-a42b-66283fd65da0" (UID: "abaf8528-a999-4343-a42b-66283fd65da0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.048168 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abaf8528-a999-4343-a42b-66283fd65da0-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.048238 4751 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-var-lock\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.048268 4751 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/abaf8528-a999-4343-a42b-66283fd65da0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.607295 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.608818 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a" exitCode=0 Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.608938 4751 scope.go:117] "RemoveContainer" containerID="5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.608915 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.611969 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"abaf8528-a999-4343-a42b-66283fd65da0","Type":"ContainerDied","Data":"d1b00e3c05347902f43ed39ea24c3c2c0be4d9421389dcd234e5ed726819f6e2"} Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.612025 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1b00e3c05347902f43ed39ea24c3c2c0be4d9421389dcd234e5ed726819f6e2" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.612109 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.632891 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.633252 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.633618 4751 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.634044 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.634939 4751 scope.go:117] "RemoveContainer" containerID="56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.637253 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.637574 4751 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.637873 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.638210 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.659672 4751 scope.go:117] "RemoveContainer" containerID="dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.700506 4751 scope.go:117] "RemoveContainer" containerID="501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.722394 4751 scope.go:117] "RemoveContainer" containerID="714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.748618 4751 scope.go:117] "RemoveContainer" containerID="2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.774772 4751 scope.go:117] "RemoveContainer" containerID="5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807" Mar 16 00:09:45 crc kubenswrapper[4751]: E0316 00:09:45.775396 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807\": container with ID starting with 5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807 not found: ID does not exist" containerID="5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.775460 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807"} err="failed to get container status \"5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807\": rpc error: code = NotFound desc = could not find container \"5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807\": container with ID starting with 5478d2ba8ffb92016115652c53297275a14830239e77206d3cbe0e669603a807 not found: ID does not exist" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.775501 4751 scope.go:117] "RemoveContainer" containerID="56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a" Mar 16 00:09:45 crc kubenswrapper[4751]: E0316 00:09:45.776261 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a\": container with ID starting with 56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a not found: ID does not exist" containerID="56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.776304 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a"} err="failed to get container status \"56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a\": rpc error: code = NotFound desc = could not find container \"56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a\": container with ID starting with 56f942772ecdf4715c8c18050990fd2ff857f1b1e05a7d78cc4fbd53f9684a1a not found: ID does not exist" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.776335 4751 scope.go:117] "RemoveContainer" containerID="dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8" Mar 16 00:09:45 crc kubenswrapper[4751]: E0316 00:09:45.776831 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8\": container with ID starting with dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8 not found: ID does not exist" containerID="dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.776884 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8"} err="failed to get container status \"dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8\": rpc error: code = NotFound desc = could not find container \"dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8\": container with ID starting with dd690ee080d49714f2aa890db0bdaf2a7d05e56c44062253db0871c22bbbf8b8 not found: ID does not exist" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.776914 4751 scope.go:117] "RemoveContainer" containerID="501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d" Mar 16 00:09:45 crc kubenswrapper[4751]: E0316 00:09:45.777528 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d\": container with ID starting with 501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d not found: ID does not exist" containerID="501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.777629 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d"} err="failed to get container status \"501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d\": rpc error: code = NotFound desc = could not find container \"501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d\": container with ID starting with 501124f5b1a96cbf3be1b53cea6bfacda84b3dd566ac055e32334b05b65a993d not found: ID does not exist" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.777719 4751 scope.go:117] "RemoveContainer" containerID="714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a" Mar 16 00:09:45 crc kubenswrapper[4751]: E0316 00:09:45.778649 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a\": container with ID starting with 714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a not found: ID does not exist" containerID="714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.778690 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a"} err="failed to get container status \"714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a\": rpc error: code = NotFound desc = could not find container \"714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a\": container with ID starting with 714353f8459be9ffdf76ccbb2f35ab7d299759a3e73058038d0079be8fabf03a not found: ID does not exist" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.778717 4751 scope.go:117] "RemoveContainer" containerID="2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057" Mar 16 00:09:45 crc kubenswrapper[4751]: E0316 00:09:45.779431 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057\": container with ID starting with 2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057 not found: ID does not exist" containerID="2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057" Mar 16 00:09:45 crc kubenswrapper[4751]: I0316 00:09:45.779518 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057"} err="failed to get container status \"2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057\": rpc error: code = NotFound desc = could not find container \"2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057\": container with ID starting with 2d0a6f48e3ed8ded56e54d98d7aac227e8d83fc93bf5c4e4d3c18230f5c33057 not found: ID does not exist" Mar 16 00:09:46 crc kubenswrapper[4751]: I0316 00:09:46.465947 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 16 00:09:49 crc kubenswrapper[4751]: E0316 00:09:49.530340 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.36:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189d29cba279d130 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-16 00:09:42.83483576 +0000 UTC m=+211.104168070,LastTimestamp:2026-03-16 00:09:42.83483576 +0000 UTC m=+211.104168070,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 16 00:09:51 crc kubenswrapper[4751]: E0316 00:09:51.321292 4751 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:51 crc kubenswrapper[4751]: E0316 00:09:51.322180 4751 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:51 crc kubenswrapper[4751]: E0316 00:09:51.323178 4751 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:51 crc kubenswrapper[4751]: E0316 00:09:51.323561 4751 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:51 crc kubenswrapper[4751]: E0316 00:09:51.323942 4751 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:51 crc kubenswrapper[4751]: I0316 00:09:51.324007 4751 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 16 00:09:51 crc kubenswrapper[4751]: E0316 00:09:51.324545 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="200ms" Mar 16 00:09:51 crc kubenswrapper[4751]: E0316 00:09:51.525963 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="400ms" Mar 16 00:09:51 crc kubenswrapper[4751]: E0316 00:09:51.927090 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="800ms" Mar 16 00:09:52 crc kubenswrapper[4751]: I0316 00:09:52.466537 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:52 crc kubenswrapper[4751]: I0316 00:09:52.467389 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:52 crc kubenswrapper[4751]: I0316 00:09:52.467861 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:52 crc kubenswrapper[4751]: E0316 00:09:52.728257 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="1.6s" Mar 16 00:09:54 crc kubenswrapper[4751]: E0316 00:09:54.329965 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="3.2s" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.454210 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.455699 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.456944 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.457777 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.475733 4751 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.476005 4751 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:09:56 crc kubenswrapper[4751]: E0316 00:09:56.476819 4751 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.477640 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:56 crc kubenswrapper[4751]: W0316 00:09:56.510816 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-5c430d2e73d2c2e306ccbbacade0c7a897883c17469b582f2bd829f74e78136e WatchSource:0}: Error finding container 5c430d2e73d2c2e306ccbbacade0c7a897883c17469b582f2bd829f74e78136e: Status 404 returned error can't find the container with id 5c430d2e73d2c2e306ccbbacade0c7a897883c17469b582f2bd829f74e78136e Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.751647 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.753344 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.753445 4751 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="eb5cc30849ffd9367bd9dba9e0eccf95ba6e364bee9b1b05ed74e4bb0cdda793" exitCode=1 Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.753564 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"eb5cc30849ffd9367bd9dba9e0eccf95ba6e364bee9b1b05ed74e4bb0cdda793"} Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.754435 4751 scope.go:117] "RemoveContainer" containerID="eb5cc30849ffd9367bd9dba9e0eccf95ba6e364bee9b1b05ed74e4bb0cdda793" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.754526 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.754608 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5c430d2e73d2c2e306ccbbacade0c7a897883c17469b582f2bd829f74e78136e"} Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.754976 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.755761 4751 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:56 crc kubenswrapper[4751]: I0316 00:09:56.756404 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:57 crc kubenswrapper[4751]: E0316 00:09:57.532137 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.36:6443: connect: connection refused" interval="6.4s" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.765439 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.768162 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.768331 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3af477f2da05186ecfb643789a22a33b455da4f477e40aeaaae58c21a7fa0c35"} Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.769647 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.770421 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.771148 4751 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b822d0ddb6adb2000c3bfae92509f61d126892d1543a2fdb0f14a1d160e30ba6" exitCode=0 Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.771212 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b822d0ddb6adb2000c3bfae92509f61d126892d1543a2fdb0f14a1d160e30ba6"} Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.771271 4751 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.771460 4751 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.771488 4751 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:09:57 crc kubenswrapper[4751]: E0316 00:09:57.771885 4751 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.772598 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.773357 4751 status_manager.go:851] "Failed to get status for pod" podUID="abaf8528-a999-4343-a42b-66283fd65da0" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.773835 4751 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.774545 4751 status_manager.go:851] "Failed to get status for pod" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" pod="openshift-controller-manager/controller-manager-d669cf5cf-5s5gd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-d669cf5cf-5s5gd\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:57 crc kubenswrapper[4751]: I0316 00:09:57.774981 4751 status_manager.go:851] "Failed to get status for pod" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" pod="openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-598cd6b6f-dbnwn\": dial tcp 38.102.83.36:6443: connect: connection refused" Mar 16 00:09:58 crc kubenswrapper[4751]: I0316 00:09:58.799970 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"51148c1c42aefa756ca85ba0c6cd64ed91883418bb3794cc21aadb8f3e623ce6"} Mar 16 00:09:58 crc kubenswrapper[4751]: I0316 00:09:58.800249 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"980e0d3ff4dd3cfcc1241e2652fe832a3f9a23c33bf46ed63de78f280673696a"} Mar 16 00:09:58 crc kubenswrapper[4751]: I0316 00:09:58.800262 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"204b8e2467549e450a1f63aaf4735167bd1608e3f561a1e8ccc9ca0b24a54188"} Mar 16 00:09:59 crc kubenswrapper[4751]: I0316 00:09:59.808833 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a25dee2013693d2dea19c3519c423ab4b5d369fd256f7a4ca12ab3338e70ed5f"} Mar 16 00:09:59 crc kubenswrapper[4751]: I0316 00:09:59.809075 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"61a0c1fe22286e1de20a6a110412a946f230bf887b8c128ff80bfc041cda927b"} Mar 16 00:09:59 crc kubenswrapper[4751]: I0316 00:09:59.809255 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:09:59 crc kubenswrapper[4751]: I0316 00:09:59.809385 4751 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:09:59 crc kubenswrapper[4751]: I0316 00:09:59.809416 4751 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:10:01 crc kubenswrapper[4751]: I0316 00:10:01.479287 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:10:01 crc kubenswrapper[4751]: I0316 00:10:01.479760 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:10:01 crc kubenswrapper[4751]: I0316 00:10:01.489047 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:10:03 crc kubenswrapper[4751]: I0316 00:10:03.071524 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:10:03 crc kubenswrapper[4751]: I0316 00:10:03.124911 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:10:03 crc kubenswrapper[4751]: I0316 00:10:03.125668 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 16 00:10:03 crc kubenswrapper[4751]: I0316 00:10:03.125754 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 16 00:10:04 crc kubenswrapper[4751]: I0316 00:10:04.826958 4751 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:10:04 crc kubenswrapper[4751]: I0316 00:10:04.914008 4751 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="793dae53-31f6-4344-9463-c15ec8bfad72" Mar 16 00:10:05 crc kubenswrapper[4751]: I0316 00:10:05.850269 4751 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:10:05 crc kubenswrapper[4751]: I0316 00:10:05.850329 4751 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:10:05 crc kubenswrapper[4751]: I0316 00:10:05.858730 4751 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="793dae53-31f6-4344-9463-c15ec8bfad72" Mar 16 00:10:05 crc kubenswrapper[4751]: I0316 00:10:05.865306 4751 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://204b8e2467549e450a1f63aaf4735167bd1608e3f561a1e8ccc9ca0b24a54188" Mar 16 00:10:05 crc kubenswrapper[4751]: I0316 00:10:05.865358 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:10:06 crc kubenswrapper[4751]: I0316 00:10:06.857347 4751 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:10:06 crc kubenswrapper[4751]: I0316 00:10:06.857401 4751 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:10:06 crc kubenswrapper[4751]: I0316 00:10:06.862616 4751 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="793dae53-31f6-4344-9463-c15ec8bfad72" Mar 16 00:10:13 crc kubenswrapper[4751]: I0316 00:10:13.130814 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:10:13 crc kubenswrapper[4751]: I0316 00:10:13.139501 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 16 00:10:14 crc kubenswrapper[4751]: I0316 00:10:14.046192 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 16 00:10:15 crc kubenswrapper[4751]: I0316 00:10:15.370744 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 16 00:10:15 crc kubenswrapper[4751]: I0316 00:10:15.718682 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.214966 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.268265 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.417825 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.530356 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.569778 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.717842 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.719821 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.855751 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.881841 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.967220 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.969016 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 16 00:10:16 crc kubenswrapper[4751]: I0316 00:10:16.983491 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.175689 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.191626 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.210511 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.284334 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.320946 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.445875 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.493549 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.516645 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.632566 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.679019 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.681185 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.691725 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.715580 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.738636 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.894826 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.901137 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 16 00:10:17 crc kubenswrapper[4751]: I0316 00:10:17.971159 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.199635 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.210875 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.339690 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.514353 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.533221 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.538978 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.561640 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.583184 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.608374 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.646043 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.700459 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.800715 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.842284 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.849141 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 16 00:10:18 crc kubenswrapper[4751]: I0316 00:10:18.985506 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.017917 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.083793 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.271934 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.281724 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.341162 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.465330 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.529727 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.541600 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.762464 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.854711 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.859328 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 16 00:10:19 crc kubenswrapper[4751]: I0316 00:10:19.947422 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.000677 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.053865 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.071049 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.103827 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.104541 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.118176 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.133236 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.167242 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.229668 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.244760 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.409746 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.551740 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.590926 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.681917 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.730971 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.732842 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.937787 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.962178 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.974743 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 16 00:10:20 crc kubenswrapper[4751]: I0316 00:10:20.977681 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.009738 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.046236 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.144304 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.195040 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.253529 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.265065 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.284998 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.351923 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.362382 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.381276 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.431011 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.438486 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.555140 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.665493 4751 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.693061 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.761552 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.875888 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.891124 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.924029 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 16 00:10:21 crc kubenswrapper[4751]: I0316 00:10:21.968031 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.022243 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.149708 4751 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.264825 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.272552 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.302273 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.351204 4751 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.352695 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.384379 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.422451 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.422492 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.527850 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.669415 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.686387 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.757505 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.772393 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 16 00:10:22 crc kubenswrapper[4751]: I0316 00:10:22.978652 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.021079 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.029959 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.044938 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.057147 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.122270 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.166812 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.179829 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.190411 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.241550 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.294411 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.330708 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.390667 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.390953 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.391604 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.392367 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.523233 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.543097 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.590562 4751 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.591592 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.591625 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.597361 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598cd6b6f-dbnwn","openshift-controller-manager/controller-manager-d669cf5cf-5s5gd","openshift-kube-apiserver/kube-apiserver-crc"] Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.597440 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p","openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48","openshift-kube-apiserver/kube-apiserver-crc"] Mar 16 00:10:23 crc kubenswrapper[4751]: E0316 00:10:23.597677 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaf8528-a999-4343-a42b-66283fd65da0" containerName="installer" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.597699 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaf8528-a999-4343-a42b-66283fd65da0" containerName="installer" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.597845 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="abaf8528-a999-4343-a42b-66283fd65da0" containerName="installer" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.598084 4751 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.598179 4751 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdd1c6a8-b99c-4864-9be1-a25ff0b789d0" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.598551 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.599302 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.608500 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.618648 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.622598 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.623215 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.630798 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.631238 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.631402 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.630089 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.632049 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.632089 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.633721 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.635402 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.641197 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.654789 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.683591 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/388be997-7133-489c-981a-df6cea7dd5f0-serving-cert\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.684706 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8ncx\" (UniqueName: \"kubernetes.io/projected/388be997-7133-489c-981a-df6cea7dd5f0-kube-api-access-s8ncx\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.684854 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/388be997-7133-489c-981a-df6cea7dd5f0-client-ca\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.685016 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60590194-b2fc-4b9d-a3c8-95806f8797d3-client-ca\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.685217 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60590194-b2fc-4b9d-a3c8-95806f8797d3-config\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.686027 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/388be997-7133-489c-981a-df6cea7dd5f0-proxy-ca-bundles\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.686207 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60590194-b2fc-4b9d-a3c8-95806f8797d3-serving-cert\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.686629 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388be997-7133-489c-981a-df6cea7dd5f0-config\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.687011 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85jjc\" (UniqueName: \"kubernetes.io/projected/60590194-b2fc-4b9d-a3c8-95806f8797d3-kube-api-access-85jjc\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.700703 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.700668845 podStartE2EDuration="19.700668845s" podCreationTimestamp="2026-03-16 00:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:10:23.7001925 +0000 UTC m=+251.969524820" watchObservedRunningTime="2026-03-16 00:10:23.700668845 +0000 UTC m=+251.970001165" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.788815 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8ncx\" (UniqueName: \"kubernetes.io/projected/388be997-7133-489c-981a-df6cea7dd5f0-kube-api-access-s8ncx\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.788957 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/388be997-7133-489c-981a-df6cea7dd5f0-client-ca\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.788993 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60590194-b2fc-4b9d-a3c8-95806f8797d3-client-ca\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.789035 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60590194-b2fc-4b9d-a3c8-95806f8797d3-config\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.789086 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/388be997-7133-489c-981a-df6cea7dd5f0-proxy-ca-bundles\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.789141 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60590194-b2fc-4b9d-a3c8-95806f8797d3-serving-cert\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.789177 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388be997-7133-489c-981a-df6cea7dd5f0-config\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.789211 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85jjc\" (UniqueName: \"kubernetes.io/projected/60590194-b2fc-4b9d-a3c8-95806f8797d3-kube-api-access-85jjc\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.789252 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/388be997-7133-489c-981a-df6cea7dd5f0-serving-cert\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.790310 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/388be997-7133-489c-981a-df6cea7dd5f0-client-ca\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.790767 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60590194-b2fc-4b9d-a3c8-95806f8797d3-client-ca\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.791124 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/388be997-7133-489c-981a-df6cea7dd5f0-proxy-ca-bundles\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.791155 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60590194-b2fc-4b9d-a3c8-95806f8797d3-config\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.792431 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/388be997-7133-489c-981a-df6cea7dd5f0-config\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.796411 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60590194-b2fc-4b9d-a3c8-95806f8797d3-serving-cert\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.809617 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/388be997-7133-489c-981a-df6cea7dd5f0-serving-cert\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.814734 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85jjc\" (UniqueName: \"kubernetes.io/projected/60590194-b2fc-4b9d-a3c8-95806f8797d3-kube-api-access-85jjc\") pod \"route-controller-manager-d7cff4579-6gx48\" (UID: \"60590194-b2fc-4b9d-a3c8-95806f8797d3\") " pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.819908 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8ncx\" (UniqueName: \"kubernetes.io/projected/388be997-7133-489c-981a-df6cea7dd5f0-kube-api-access-s8ncx\") pod \"controller-manager-6bcd5bbbfb-fkx9p\" (UID: \"388be997-7133-489c-981a-df6cea7dd5f0\") " pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.833496 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.929732 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.943497 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.945230 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.953835 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 16 00:10:23 crc kubenswrapper[4751]: I0316 00:10:23.956153 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.023327 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.096719 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.130274 4751 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.401300 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.460653 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="642903d1-a399-42c9-94f3-97eb5c4ca96e" path="/var/lib/kubelet/pods/642903d1-a399-42c9-94f3-97eb5c4ca96e/volumes" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.461350 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b469eae7-c974-43b0-bf74-706cc2bf62aa" path="/var/lib/kubelet/pods/b469eae7-c974-43b0-bf74-706cc2bf62aa/volumes" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.466361 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.516295 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.529228 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.567703 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.610208 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.644216 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.661147 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.664580 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.733242 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.737048 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.922880 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 16 00:10:24 crc kubenswrapper[4751]: I0316 00:10:24.967931 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.199005 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.208910 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.214010 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.240077 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.267176 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.268051 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.386719 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.397314 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.425138 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.446476 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.608359 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.643570 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.717257 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.721176 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560330-lnjc4"] Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.722178 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560330-lnjc4" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.724662 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.724730 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.724743 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.807759 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.809637 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.858334 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.914056 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfvnt\" (UniqueName: \"kubernetes.io/projected/f4b44b74-5471-4589-83da-e40698b419f0-kube-api-access-cfvnt\") pod \"auto-csr-approver-29560330-lnjc4\" (UID: \"f4b44b74-5471-4589-83da-e40698b419f0\") " pod="openshift-infra/auto-csr-approver-29560330-lnjc4" Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.924649 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560330-lnjc4"] Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.930253 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p"] Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.958324 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48"] Mar 16 00:10:25 crc kubenswrapper[4751]: I0316 00:10:25.965266 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.016240 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfvnt\" (UniqueName: \"kubernetes.io/projected/f4b44b74-5471-4589-83da-e40698b419f0-kube-api-access-cfvnt\") pod \"auto-csr-approver-29560330-lnjc4\" (UID: \"f4b44b74-5471-4589-83da-e40698b419f0\") " pod="openshift-infra/auto-csr-approver-29560330-lnjc4" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.035674 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfvnt\" (UniqueName: \"kubernetes.io/projected/f4b44b74-5471-4589-83da-e40698b419f0-kube-api-access-cfvnt\") pod \"auto-csr-approver-29560330-lnjc4\" (UID: \"f4b44b74-5471-4589-83da-e40698b419f0\") " pod="openshift-infra/auto-csr-approver-29560330-lnjc4" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.047737 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560330-lnjc4" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.159138 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.174436 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.223484 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.241945 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.352068 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.375836 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.447723 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.614990 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.689356 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.738629 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p"] Mar 16 00:10:26 crc kubenswrapper[4751]: W0316 00:10:26.755415 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod388be997_7133_489c_981a_df6cea7dd5f0.slice/crio-d3f73e32633b9a75f07e940562db8bb2fb13cd670a361a74db69859fa834e963 WatchSource:0}: Error finding container d3f73e32633b9a75f07e940562db8bb2fb13cd670a361a74db69859fa834e963: Status 404 returned error can't find the container with id d3f73e32633b9a75f07e940562db8bb2fb13cd670a361a74db69859fa834e963 Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.886248 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560330-lnjc4"] Mar 16 00:10:26 crc kubenswrapper[4751]: W0316 00:10:26.898555 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4b44b74_5471_4589_83da_e40698b419f0.slice/crio-3551cd8a6c1ea0eb799d60304f30c0952b512790ff4b6ff1a81183b7a3a609f2 WatchSource:0}: Error finding container 3551cd8a6c1ea0eb799d60304f30c0952b512790ff4b6ff1a81183b7a3a609f2: Status 404 returned error can't find the container with id 3551cd8a6c1ea0eb799d60304f30c0952b512790ff4b6ff1a81183b7a3a609f2 Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.909652 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.933573 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.991476 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48"] Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.992446 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" event={"ID":"388be997-7133-489c-981a-df6cea7dd5f0","Type":"ContainerStarted","Data":"f46dfb66fd4798785eb43d7722f780661fa79aba2a5f365ebcce1a6277c760d5"} Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.992505 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" event={"ID":"388be997-7133-489c-981a-df6cea7dd5f0","Type":"ContainerStarted","Data":"d3f73e32633b9a75f07e940562db8bb2fb13cd670a361a74db69859fa834e963"} Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.992617 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.993946 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560330-lnjc4" event={"ID":"f4b44b74-5471-4589-83da-e40698b419f0","Type":"ContainerStarted","Data":"3551cd8a6c1ea0eb799d60304f30c0952b512790ff4b6ff1a81183b7a3a609f2"} Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.994014 4751 patch_prober.go:28] interesting pod/controller-manager-6bcd5bbbfb-fkx9p container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" start-of-body= Mar 16 00:10:26 crc kubenswrapper[4751]: I0316 00:10:26.994050 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" podUID="388be997-7133-489c-981a-df6cea7dd5f0" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" Mar 16 00:10:27 crc kubenswrapper[4751]: W0316 00:10:27.001464 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60590194_b2fc_4b9d_a3c8_95806f8797d3.slice/crio-12492bf1a294b0a059b8f7e31b6091c3cf4d3ac3b07f0560b1db945787ab1516 WatchSource:0}: Error finding container 12492bf1a294b0a059b8f7e31b6091c3cf4d3ac3b07f0560b1db945787ab1516: Status 404 returned error can't find the container with id 12492bf1a294b0a059b8f7e31b6091c3cf4d3ac3b07f0560b1db945787ab1516 Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.009965 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" podStartSLOduration=46.009949121 podStartE2EDuration="46.009949121s" podCreationTimestamp="2026-03-16 00:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:10:27.009394934 +0000 UTC m=+255.278727224" watchObservedRunningTime="2026-03-16 00:10:27.009949121 +0000 UTC m=+255.279281411" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.062911 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.075672 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.130379 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.246263 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.364607 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.371423 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.380308 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.410372 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.447154 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.510016 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.519631 4751 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.519845 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://3066c873ba83a58e27cc3b8b9c8be7a3ab903a30be5e955869b15921476ef076" gracePeriod=5 Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.566118 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.641591 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.667015 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.764452 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.854607 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.854679 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.869823 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.870594 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.933464 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.966165 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 16 00:10:27 crc kubenswrapper[4751]: I0316 00:10:27.967640 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.000479 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" event={"ID":"60590194-b2fc-4b9d-a3c8-95806f8797d3","Type":"ContainerStarted","Data":"aa2ebbce67e0631ac495fef87093a459f85cb9baaaa346140ba8bab16d04e6a6"} Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.000534 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" event={"ID":"60590194-b2fc-4b9d-a3c8-95806f8797d3","Type":"ContainerStarted","Data":"12492bf1a294b0a059b8f7e31b6091c3cf4d3ac3b07f0560b1db945787ab1516"} Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.005735 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6bcd5bbbfb-fkx9p" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.024352 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" podStartSLOduration=47.024332734 podStartE2EDuration="47.024332734s" podCreationTimestamp="2026-03-16 00:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:10:28.020662884 +0000 UTC m=+256.289995174" watchObservedRunningTime="2026-03-16 00:10:28.024332734 +0000 UTC m=+256.293665024" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.071446 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.093313 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.233200 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.233256 4751 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.264982 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.395207 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.444338 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.607640 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.619193 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 16 00:10:28 crc kubenswrapper[4751]: I0316 00:10:28.662508 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 16 00:10:29 crc kubenswrapper[4751]: I0316 00:10:29.005869 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:29 crc kubenswrapper[4751]: I0316 00:10:29.010398 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-d7cff4579-6gx48" Mar 16 00:10:29 crc kubenswrapper[4751]: I0316 00:10:29.029882 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 16 00:10:29 crc kubenswrapper[4751]: I0316 00:10:29.141169 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 16 00:10:29 crc kubenswrapper[4751]: I0316 00:10:29.710241 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 16 00:10:29 crc kubenswrapper[4751]: I0316 00:10:29.722989 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 16 00:10:29 crc kubenswrapper[4751]: I0316 00:10:29.894588 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 16 00:10:29 crc kubenswrapper[4751]: I0316 00:10:29.962154 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 16 00:10:29 crc kubenswrapper[4751]: I0316 00:10:29.989401 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 16 00:10:30 crc kubenswrapper[4751]: I0316 00:10:30.011452 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b44b74-5471-4589-83da-e40698b419f0" containerID="ea3548a4dabf550b17c81a439a332770418cf213f78f87e6ac23ec95f317f195" exitCode=0 Mar 16 00:10:30 crc kubenswrapper[4751]: I0316 00:10:30.011818 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560330-lnjc4" event={"ID":"f4b44b74-5471-4589-83da-e40698b419f0","Type":"ContainerDied","Data":"ea3548a4dabf550b17c81a439a332770418cf213f78f87e6ac23ec95f317f195"} Mar 16 00:10:30 crc kubenswrapper[4751]: I0316 00:10:30.060565 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 16 00:10:30 crc kubenswrapper[4751]: I0316 00:10:30.158343 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 16 00:10:30 crc kubenswrapper[4751]: I0316 00:10:30.260578 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 16 00:10:30 crc kubenswrapper[4751]: I0316 00:10:30.669735 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 16 00:10:31 crc kubenswrapper[4751]: I0316 00:10:31.409747 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560330-lnjc4" Mar 16 00:10:31 crc kubenswrapper[4751]: I0316 00:10:31.598973 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfvnt\" (UniqueName: \"kubernetes.io/projected/f4b44b74-5471-4589-83da-e40698b419f0-kube-api-access-cfvnt\") pod \"f4b44b74-5471-4589-83da-e40698b419f0\" (UID: \"f4b44b74-5471-4589-83da-e40698b419f0\") " Mar 16 00:10:31 crc kubenswrapper[4751]: I0316 00:10:31.606203 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4b44b74-5471-4589-83da-e40698b419f0-kube-api-access-cfvnt" (OuterVolumeSpecName: "kube-api-access-cfvnt") pod "f4b44b74-5471-4589-83da-e40698b419f0" (UID: "f4b44b74-5471-4589-83da-e40698b419f0"). InnerVolumeSpecName "kube-api-access-cfvnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:10:31 crc kubenswrapper[4751]: I0316 00:10:31.699713 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfvnt\" (UniqueName: \"kubernetes.io/projected/f4b44b74-5471-4589-83da-e40698b419f0-kube-api-access-cfvnt\") on node \"crc\" DevicePath \"\"" Mar 16 00:10:32 crc kubenswrapper[4751]: I0316 00:10:32.025615 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560330-lnjc4" event={"ID":"f4b44b74-5471-4589-83da-e40698b419f0","Type":"ContainerDied","Data":"3551cd8a6c1ea0eb799d60304f30c0952b512790ff4b6ff1a81183b7a3a609f2"} Mar 16 00:10:32 crc kubenswrapper[4751]: I0316 00:10:32.025683 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3551cd8a6c1ea0eb799d60304f30c0952b512790ff4b6ff1a81183b7a3a609f2" Mar 16 00:10:32 crc kubenswrapper[4751]: I0316 00:10:32.025707 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560330-lnjc4" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.033853 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.033922 4751 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="3066c873ba83a58e27cc3b8b9c8be7a3ab903a30be5e955869b15921476ef076" exitCode=137 Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.110927 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.111507 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.222419 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.222848 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.222924 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.222855 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.223000 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.223030 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.223073 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.223207 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.223257 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.223450 4751 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.223517 4751 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.223540 4751 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.223560 4751 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.234596 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:10:33 crc kubenswrapper[4751]: I0316 00:10:33.325394 4751 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 16 00:10:34 crc kubenswrapper[4751]: I0316 00:10:34.041616 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 16 00:10:34 crc kubenswrapper[4751]: I0316 00:10:34.041714 4751 scope.go:117] "RemoveContainer" containerID="3066c873ba83a58e27cc3b8b9c8be7a3ab903a30be5e955869b15921476ef076" Mar 16 00:10:34 crc kubenswrapper[4751]: I0316 00:10:34.041755 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 16 00:10:34 crc kubenswrapper[4751]: I0316 00:10:34.467190 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 16 00:10:37 crc kubenswrapper[4751]: I0316 00:10:37.566534 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 16 00:10:40 crc kubenswrapper[4751]: I0316 00:10:40.350763 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 16 00:10:43 crc kubenswrapper[4751]: I0316 00:10:43.557017 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 16 00:10:47 crc kubenswrapper[4751]: I0316 00:10:47.213747 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 16 00:10:47 crc kubenswrapper[4751]: I0316 00:10:47.672900 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 16 00:10:47 crc kubenswrapper[4751]: I0316 00:10:47.749763 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 16 00:10:49 crc kubenswrapper[4751]: I0316 00:10:49.124150 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 16 00:10:49 crc kubenswrapper[4751]: I0316 00:10:49.375035 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 16 00:10:49 crc kubenswrapper[4751]: I0316 00:10:49.492694 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 16 00:10:51 crc kubenswrapper[4751]: I0316 00:10:51.035169 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 16 00:10:51 crc kubenswrapper[4751]: I0316 00:10:51.976705 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 16 00:10:52 crc kubenswrapper[4751]: I0316 00:10:52.175304 4751 generic.go:334] "Generic (PLEG): container finished" podID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerID="bc4a634b7a61211753a19f21449ebaa9564426678f0caac161542bf34aafd110" exitCode=0 Mar 16 00:10:52 crc kubenswrapper[4751]: I0316 00:10:52.175378 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" event={"ID":"69dee6c1-1c2f-4d49-8955-ca4226b4306e","Type":"ContainerDied","Data":"bc4a634b7a61211753a19f21449ebaa9564426678f0caac161542bf34aafd110"} Mar 16 00:10:52 crc kubenswrapper[4751]: I0316 00:10:52.176095 4751 scope.go:117] "RemoveContainer" containerID="bc4a634b7a61211753a19f21449ebaa9564426678f0caac161542bf34aafd110" Mar 16 00:10:53 crc kubenswrapper[4751]: I0316 00:10:53.184749 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" event={"ID":"69dee6c1-1c2f-4d49-8955-ca4226b4306e","Type":"ContainerStarted","Data":"9ba145bd0d82918f5fcf95cb700001241529e0f4b695b8cead72e61eef415a21"} Mar 16 00:10:53 crc kubenswrapper[4751]: I0316 00:10:53.185570 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:10:53 crc kubenswrapper[4751]: I0316 00:10:53.187847 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:10:54 crc kubenswrapper[4751]: I0316 00:10:54.856061 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 16 00:10:55 crc kubenswrapper[4751]: I0316 00:10:55.823877 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 16 00:10:56 crc kubenswrapper[4751]: I0316 00:10:56.562961 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 16 00:10:57 crc kubenswrapper[4751]: I0316 00:10:57.855391 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:10:57 crc kubenswrapper[4751]: I0316 00:10:57.855567 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:10:59 crc kubenswrapper[4751]: I0316 00:10:59.403516 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 16 00:11:04 crc kubenswrapper[4751]: I0316 00:11:04.643179 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 16 00:11:10 crc kubenswrapper[4751]: I0316 00:11:10.939752 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-p8ptf"] Mar 16 00:11:10 crc kubenswrapper[4751]: E0316 00:11:10.940326 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 16 00:11:10 crc kubenswrapper[4751]: I0316 00:11:10.940340 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 16 00:11:10 crc kubenswrapper[4751]: E0316 00:11:10.940353 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b44b74-5471-4589-83da-e40698b419f0" containerName="oc" Mar 16 00:11:10 crc kubenswrapper[4751]: I0316 00:11:10.940361 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b44b74-5471-4589-83da-e40698b419f0" containerName="oc" Mar 16 00:11:10 crc kubenswrapper[4751]: I0316 00:11:10.940497 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 16 00:11:10 crc kubenswrapper[4751]: I0316 00:11:10.940512 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b44b74-5471-4589-83da-e40698b419f0" containerName="oc" Mar 16 00:11:10 crc kubenswrapper[4751]: I0316 00:11:10.940955 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:10 crc kubenswrapper[4751]: I0316 00:11:10.965676 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-p8ptf"] Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.099848 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa87974-0d03-4ce3-8271-77de98ba99cf-registry-tls\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.099905 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3aa87974-0d03-4ce3-8271-77de98ba99cf-installation-pull-secrets\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.099955 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3aa87974-0d03-4ce3-8271-77de98ba99cf-bound-sa-token\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.100094 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2nt8\" (UniqueName: \"kubernetes.io/projected/3aa87974-0d03-4ce3-8271-77de98ba99cf-kube-api-access-r2nt8\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.100247 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3aa87974-0d03-4ce3-8271-77de98ba99cf-trusted-ca\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.100312 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3aa87974-0d03-4ce3-8271-77de98ba99cf-ca-trust-extracted\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.100353 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.100396 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3aa87974-0d03-4ce3-8271-77de98ba99cf-registry-certificates\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.130132 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.201870 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa87974-0d03-4ce3-8271-77de98ba99cf-registry-tls\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.201934 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3aa87974-0d03-4ce3-8271-77de98ba99cf-installation-pull-secrets\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.201986 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3aa87974-0d03-4ce3-8271-77de98ba99cf-bound-sa-token\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.202024 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2nt8\" (UniqueName: \"kubernetes.io/projected/3aa87974-0d03-4ce3-8271-77de98ba99cf-kube-api-access-r2nt8\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.202072 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3aa87974-0d03-4ce3-8271-77de98ba99cf-trusted-ca\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.202099 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3aa87974-0d03-4ce3-8271-77de98ba99cf-ca-trust-extracted\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.202167 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3aa87974-0d03-4ce3-8271-77de98ba99cf-registry-certificates\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.202806 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3aa87974-0d03-4ce3-8271-77de98ba99cf-ca-trust-extracted\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.203683 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3aa87974-0d03-4ce3-8271-77de98ba99cf-trusted-ca\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.203787 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3aa87974-0d03-4ce3-8271-77de98ba99cf-registry-certificates\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.207879 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3aa87974-0d03-4ce3-8271-77de98ba99cf-registry-tls\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.210653 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3aa87974-0d03-4ce3-8271-77de98ba99cf-installation-pull-secrets\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.220805 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2nt8\" (UniqueName: \"kubernetes.io/projected/3aa87974-0d03-4ce3-8271-77de98ba99cf-kube-api-access-r2nt8\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.223732 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3aa87974-0d03-4ce3-8271-77de98ba99cf-bound-sa-token\") pod \"image-registry-66df7c8f76-p8ptf\" (UID: \"3aa87974-0d03-4ce3-8271-77de98ba99cf\") " pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.259529 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:11 crc kubenswrapper[4751]: I0316 00:11:11.687415 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-p8ptf"] Mar 16 00:11:11 crc kubenswrapper[4751]: W0316 00:11:11.693613 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3aa87974_0d03_4ce3_8271_77de98ba99cf.slice/crio-67bd854aed4d3cd53c10f9c345c06f7559b10fc2e31d8a774c75dee13abf686b WatchSource:0}: Error finding container 67bd854aed4d3cd53c10f9c345c06f7559b10fc2e31d8a774c75dee13abf686b: Status 404 returned error can't find the container with id 67bd854aed4d3cd53c10f9c345c06f7559b10fc2e31d8a774c75dee13abf686b Mar 16 00:11:12 crc kubenswrapper[4751]: I0316 00:11:12.306310 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" event={"ID":"3aa87974-0d03-4ce3-8271-77de98ba99cf","Type":"ContainerStarted","Data":"77a14baad2b6aa3b99e12bb6a9b9e7ae0d8de43934528921d7513ddaad538a18"} Mar 16 00:11:12 crc kubenswrapper[4751]: I0316 00:11:12.306711 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" event={"ID":"3aa87974-0d03-4ce3-8271-77de98ba99cf","Type":"ContainerStarted","Data":"67bd854aed4d3cd53c10f9c345c06f7559b10fc2e31d8a774c75dee13abf686b"} Mar 16 00:11:12 crc kubenswrapper[4751]: I0316 00:11:12.306739 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:12 crc kubenswrapper[4751]: I0316 00:11:12.342944 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" podStartSLOduration=2.342920778 podStartE2EDuration="2.342920778s" podCreationTimestamp="2026-03-16 00:11:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:11:12.342719592 +0000 UTC m=+300.612051922" watchObservedRunningTime="2026-03-16 00:11:12.342920778 +0000 UTC m=+300.612253098" Mar 16 00:11:27 crc kubenswrapper[4751]: I0316 00:11:27.855260 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:11:27 crc kubenswrapper[4751]: I0316 00:11:27.855939 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:11:27 crc kubenswrapper[4751]: I0316 00:11:27.856008 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:11:27 crc kubenswrapper[4751]: I0316 00:11:27.856903 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb29dfbd6c43f45308629c345a9e5bfbd8b4e1c106c34f4c658c32496548540a"} pod="openshift-machine-config-operator/machine-config-daemon-6fccz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 16 00:11:27 crc kubenswrapper[4751]: I0316 00:11:27.857031 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" containerID="cri-o://eb29dfbd6c43f45308629c345a9e5bfbd8b4e1c106c34f4c658c32496548540a" gracePeriod=600 Mar 16 00:11:28 crc kubenswrapper[4751]: I0316 00:11:28.416904 4751 generic.go:334] "Generic (PLEG): container finished" podID="25429719-d18d-4909-8455-701d09423c36" containerID="eb29dfbd6c43f45308629c345a9e5bfbd8b4e1c106c34f4c658c32496548540a" exitCode=0 Mar 16 00:11:28 crc kubenswrapper[4751]: I0316 00:11:28.417028 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerDied","Data":"eb29dfbd6c43f45308629c345a9e5bfbd8b4e1c106c34f4c658c32496548540a"} Mar 16 00:11:29 crc kubenswrapper[4751]: I0316 00:11:29.428619 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerStarted","Data":"36270f1e57b171978cc55e8e0d2e8251ea29ff34e7dce546d8c65c5b8b799266"} Mar 16 00:11:31 crc kubenswrapper[4751]: I0316 00:11:31.268751 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-p8ptf" Mar 16 00:11:31 crc kubenswrapper[4751]: I0316 00:11:31.337077 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zd5fn"] Mar 16 00:11:56 crc kubenswrapper[4751]: I0316 00:11:56.400947 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" podUID="b695c2c9-0b76-4e93-a427-76e157f001fb" containerName="registry" containerID="cri-o://eaf8445d4fafcb92547b8b51320f812a4fc53e42ed2ad3959af9061004359b23" gracePeriod=30 Mar 16 00:11:56 crc kubenswrapper[4751]: I0316 00:11:56.623503 4751 generic.go:334] "Generic (PLEG): container finished" podID="b695c2c9-0b76-4e93-a427-76e157f001fb" containerID="eaf8445d4fafcb92547b8b51320f812a4fc53e42ed2ad3959af9061004359b23" exitCode=0 Mar 16 00:11:56 crc kubenswrapper[4751]: I0316 00:11:56.623571 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" event={"ID":"b695c2c9-0b76-4e93-a427-76e157f001fb","Type":"ContainerDied","Data":"eaf8445d4fafcb92547b8b51320f812a4fc53e42ed2ad3959af9061004359b23"} Mar 16 00:11:56 crc kubenswrapper[4751]: I0316 00:11:56.940216 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.077431 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-bound-sa-token\") pod \"b695c2c9-0b76-4e93-a427-76e157f001fb\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.077507 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn2cx\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-kube-api-access-dn2cx\") pod \"b695c2c9-0b76-4e93-a427-76e157f001fb\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.077553 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-trusted-ca\") pod \"b695c2c9-0b76-4e93-a427-76e157f001fb\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.077877 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"b695c2c9-0b76-4e93-a427-76e157f001fb\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.077976 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b695c2c9-0b76-4e93-a427-76e157f001fb-installation-pull-secrets\") pod \"b695c2c9-0b76-4e93-a427-76e157f001fb\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.078044 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-tls\") pod \"b695c2c9-0b76-4e93-a427-76e157f001fb\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.078147 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b695c2c9-0b76-4e93-a427-76e157f001fb-ca-trust-extracted\") pod \"b695c2c9-0b76-4e93-a427-76e157f001fb\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.078199 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-certificates\") pod \"b695c2c9-0b76-4e93-a427-76e157f001fb\" (UID: \"b695c2c9-0b76-4e93-a427-76e157f001fb\") " Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.079394 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "b695c2c9-0b76-4e93-a427-76e157f001fb" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.079613 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "b695c2c9-0b76-4e93-a427-76e157f001fb" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.086206 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b695c2c9-0b76-4e93-a427-76e157f001fb-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "b695c2c9-0b76-4e93-a427-76e157f001fb" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.086292 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-kube-api-access-dn2cx" (OuterVolumeSpecName: "kube-api-access-dn2cx") pod "b695c2c9-0b76-4e93-a427-76e157f001fb" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb"). InnerVolumeSpecName "kube-api-access-dn2cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.086802 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "b695c2c9-0b76-4e93-a427-76e157f001fb" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.093078 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "b695c2c9-0b76-4e93-a427-76e157f001fb" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.126374 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b695c2c9-0b76-4e93-a427-76e157f001fb-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "b695c2c9-0b76-4e93-a427-76e157f001fb" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.139902 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "b695c2c9-0b76-4e93-a427-76e157f001fb" (UID: "b695c2c9-0b76-4e93-a427-76e157f001fb"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.180226 4751 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.180271 4751 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b695c2c9-0b76-4e93-a427-76e157f001fb-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.180286 4751 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.180299 4751 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.180310 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn2cx\" (UniqueName: \"kubernetes.io/projected/b695c2c9-0b76-4e93-a427-76e157f001fb-kube-api-access-dn2cx\") on node \"crc\" DevicePath \"\"" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.180320 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b695c2c9-0b76-4e93-a427-76e157f001fb-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.180331 4751 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b695c2c9-0b76-4e93-a427-76e157f001fb-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.633178 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" event={"ID":"b695c2c9-0b76-4e93-a427-76e157f001fb","Type":"ContainerDied","Data":"483d3baff0187af0e9facbc9c29a244e7a7e1e6bfbace6d8faa206432bbbdad9"} Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.633262 4751 scope.go:117] "RemoveContainer" containerID="eaf8445d4fafcb92547b8b51320f812a4fc53e42ed2ad3959af9061004359b23" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.634286 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zd5fn" Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.702506 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zd5fn"] Mar 16 00:11:57 crc kubenswrapper[4751]: I0316 00:11:57.706907 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zd5fn"] Mar 16 00:11:58 crc kubenswrapper[4751]: I0316 00:11:58.467359 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b695c2c9-0b76-4e93-a427-76e157f001fb" path="/var/lib/kubelet/pods/b695c2c9-0b76-4e93-a427-76e157f001fb/volumes" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.149000 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560332-vtxlh"] Mar 16 00:12:00 crc kubenswrapper[4751]: E0316 00:12:00.150695 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b695c2c9-0b76-4e93-a427-76e157f001fb" containerName="registry" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.150760 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="b695c2c9-0b76-4e93-a427-76e157f001fb" containerName="registry" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.151036 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="b695c2c9-0b76-4e93-a427-76e157f001fb" containerName="registry" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.151753 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560332-vtxlh" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.155320 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.155384 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.158098 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.167020 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560332-vtxlh"] Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.330513 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb5qg\" (UniqueName: \"kubernetes.io/projected/44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd-kube-api-access-cb5qg\") pod \"auto-csr-approver-29560332-vtxlh\" (UID: \"44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd\") " pod="openshift-infra/auto-csr-approver-29560332-vtxlh" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.431853 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb5qg\" (UniqueName: \"kubernetes.io/projected/44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd-kube-api-access-cb5qg\") pod \"auto-csr-approver-29560332-vtxlh\" (UID: \"44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd\") " pod="openshift-infra/auto-csr-approver-29560332-vtxlh" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.466009 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb5qg\" (UniqueName: \"kubernetes.io/projected/44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd-kube-api-access-cb5qg\") pod \"auto-csr-approver-29560332-vtxlh\" (UID: \"44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd\") " pod="openshift-infra/auto-csr-approver-29560332-vtxlh" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.488429 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560332-vtxlh" Mar 16 00:12:00 crc kubenswrapper[4751]: I0316 00:12:00.736427 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560332-vtxlh"] Mar 16 00:12:01 crc kubenswrapper[4751]: I0316 00:12:01.671194 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560332-vtxlh" event={"ID":"44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd","Type":"ContainerStarted","Data":"50cd98db6194b44f997ae512eb4efb239008c8e89c38cc0c23585829f52dba68"} Mar 16 00:12:02 crc kubenswrapper[4751]: I0316 00:12:02.679252 4751 generic.go:334] "Generic (PLEG): container finished" podID="44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd" containerID="243ee17a8b9df926b7fee9bc4ee3b69249c33eb51fe3c27452a104ef90b678b2" exitCode=0 Mar 16 00:12:02 crc kubenswrapper[4751]: I0316 00:12:02.679299 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560332-vtxlh" event={"ID":"44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd","Type":"ContainerDied","Data":"243ee17a8b9df926b7fee9bc4ee3b69249c33eb51fe3c27452a104ef90b678b2"} Mar 16 00:12:04 crc kubenswrapper[4751]: I0316 00:12:04.008721 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560332-vtxlh" Mar 16 00:12:04 crc kubenswrapper[4751]: I0316 00:12:04.085334 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb5qg\" (UniqueName: \"kubernetes.io/projected/44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd-kube-api-access-cb5qg\") pod \"44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd\" (UID: \"44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd\") " Mar 16 00:12:04 crc kubenswrapper[4751]: I0316 00:12:04.093932 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd-kube-api-access-cb5qg" (OuterVolumeSpecName: "kube-api-access-cb5qg") pod "44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd" (UID: "44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd"). InnerVolumeSpecName "kube-api-access-cb5qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:12:04 crc kubenswrapper[4751]: I0316 00:12:04.187336 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb5qg\" (UniqueName: \"kubernetes.io/projected/44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd-kube-api-access-cb5qg\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:04 crc kubenswrapper[4751]: I0316 00:12:04.697407 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560332-vtxlh" event={"ID":"44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd","Type":"ContainerDied","Data":"50cd98db6194b44f997ae512eb4efb239008c8e89c38cc0c23585829f52dba68"} Mar 16 00:12:04 crc kubenswrapper[4751]: I0316 00:12:04.697879 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50cd98db6194b44f997ae512eb4efb239008c8e89c38cc0c23585829f52dba68" Mar 16 00:12:04 crc kubenswrapper[4751]: I0316 00:12:04.697577 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560332-vtxlh" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.598009 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqzbn"] Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.598995 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mqzbn" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerName="registry-server" containerID="cri-o://f64239b9f399ae12b962432d2a8d18f2c66c15c8e2cbdf468060f94c05b1103b" gracePeriod=30 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.602763 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8kc6w"] Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.603660 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8kc6w" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerName="registry-server" containerID="cri-o://7c86bd0dab7affbe02238e2fc8a5765d49551ed62988d987b925aa2e3f691f7a" gracePeriod=30 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.615163 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sph8v"] Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.615407 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerName="marketplace-operator" containerID="cri-o://9ba145bd0d82918f5fcf95cb700001241529e0f4b695b8cead72e61eef415a21" gracePeriod=30 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.644360 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc9vc"] Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.644679 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pc9vc" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerName="registry-server" containerID="cri-o://7932415c0e6d1902552b5f52c21327885c0b12df275ce4e58d0a0259b3a42414" gracePeriod=30 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.651068 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tlrgw"] Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.651311 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tlrgw" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerName="registry-server" containerID="cri-o://30ba8ff4e31de5142bcf76a3e58621e7063f410fb08e9cc5e4c48a96c58c9ac7" gracePeriod=30 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.656065 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ksmlj"] Mar 16 00:12:22 crc kubenswrapper[4751]: E0316 00:12:22.656718 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd" containerName="oc" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.659530 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd" containerName="oc" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.659980 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd" containerName="oc" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.660541 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.666824 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ksmlj"] Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.761607 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9674dab9-cf73-46a0-8077-1ec770c4f73c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ksmlj\" (UID: \"9674dab9-cf73-46a0-8077-1ec770c4f73c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.761648 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9674dab9-cf73-46a0-8077-1ec770c4f73c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ksmlj\" (UID: \"9674dab9-cf73-46a0-8077-1ec770c4f73c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.761711 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6lp7\" (UniqueName: \"kubernetes.io/projected/9674dab9-cf73-46a0-8077-1ec770c4f73c-kube-api-access-j6lp7\") pod \"marketplace-operator-79b997595-ksmlj\" (UID: \"9674dab9-cf73-46a0-8077-1ec770c4f73c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.853342 4751 generic.go:334] "Generic (PLEG): container finished" podID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerID="7932415c0e6d1902552b5f52c21327885c0b12df275ce4e58d0a0259b3a42414" exitCode=0 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.853411 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc9vc" event={"ID":"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4","Type":"ContainerDied","Data":"7932415c0e6d1902552b5f52c21327885c0b12df275ce4e58d0a0259b3a42414"} Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.855597 4751 generic.go:334] "Generic (PLEG): container finished" podID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerID="30ba8ff4e31de5142bcf76a3e58621e7063f410fb08e9cc5e4c48a96c58c9ac7" exitCode=0 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.855649 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlrgw" event={"ID":"8f07095b-d178-4ef1-9269-b47e33d71ef3","Type":"ContainerDied","Data":"30ba8ff4e31de5142bcf76a3e58621e7063f410fb08e9cc5e4c48a96c58c9ac7"} Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.857766 4751 generic.go:334] "Generic (PLEG): container finished" podID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerID="9ba145bd0d82918f5fcf95cb700001241529e0f4b695b8cead72e61eef415a21" exitCode=0 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.857816 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" event={"ID":"69dee6c1-1c2f-4d49-8955-ca4226b4306e","Type":"ContainerDied","Data":"9ba145bd0d82918f5fcf95cb700001241529e0f4b695b8cead72e61eef415a21"} Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.857843 4751 scope.go:117] "RemoveContainer" containerID="bc4a634b7a61211753a19f21449ebaa9564426678f0caac161542bf34aafd110" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.860712 4751 generic.go:334] "Generic (PLEG): container finished" podID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerID="7c86bd0dab7affbe02238e2fc8a5765d49551ed62988d987b925aa2e3f691f7a" exitCode=0 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.860759 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kc6w" event={"ID":"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df","Type":"ContainerDied","Data":"7c86bd0dab7affbe02238e2fc8a5765d49551ed62988d987b925aa2e3f691f7a"} Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.862427 4751 generic.go:334] "Generic (PLEG): container finished" podID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerID="f64239b9f399ae12b962432d2a8d18f2c66c15c8e2cbdf468060f94c05b1103b" exitCode=0 Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.862450 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqzbn" event={"ID":"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c","Type":"ContainerDied","Data":"f64239b9f399ae12b962432d2a8d18f2c66c15c8e2cbdf468060f94c05b1103b"} Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.862744 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9674dab9-cf73-46a0-8077-1ec770c4f73c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ksmlj\" (UID: \"9674dab9-cf73-46a0-8077-1ec770c4f73c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.862771 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9674dab9-cf73-46a0-8077-1ec770c4f73c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ksmlj\" (UID: \"9674dab9-cf73-46a0-8077-1ec770c4f73c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.862836 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6lp7\" (UniqueName: \"kubernetes.io/projected/9674dab9-cf73-46a0-8077-1ec770c4f73c-kube-api-access-j6lp7\") pod \"marketplace-operator-79b997595-ksmlj\" (UID: \"9674dab9-cf73-46a0-8077-1ec770c4f73c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.864859 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9674dab9-cf73-46a0-8077-1ec770c4f73c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ksmlj\" (UID: \"9674dab9-cf73-46a0-8077-1ec770c4f73c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.872988 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9674dab9-cf73-46a0-8077-1ec770c4f73c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ksmlj\" (UID: \"9674dab9-cf73-46a0-8077-1ec770c4f73c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:22 crc kubenswrapper[4751]: I0316 00:12:22.880319 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6lp7\" (UniqueName: \"kubernetes.io/projected/9674dab9-cf73-46a0-8077-1ec770c4f73c-kube-api-access-j6lp7\") pod \"marketplace-operator-79b997595-ksmlj\" (UID: \"9674dab9-cf73-46a0-8077-1ec770c4f73c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.039651 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.040224 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.094346 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.104857 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.123887 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.129135 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.166350 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-catalog-content\") pod \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.166402 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6kdp\" (UniqueName: \"kubernetes.io/projected/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-kube-api-access-f6kdp\") pod \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.166453 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-catalog-content\") pod \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.166487 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-utilities\") pod \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.166547 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljkkl\" (UniqueName: \"kubernetes.io/projected/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-kube-api-access-ljkkl\") pod \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\" (UID: \"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.166565 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-utilities\") pod \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\" (UID: \"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.168290 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-utilities" (OuterVolumeSpecName: "utilities") pod "8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" (UID: "8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.169372 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-utilities" (OuterVolumeSpecName: "utilities") pod "1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" (UID: "1d8a4b98-9f6d-4bfa-abba-ee21c09f51df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.171290 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-kube-api-access-f6kdp" (OuterVolumeSpecName: "kube-api-access-f6kdp") pod "1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" (UID: "1d8a4b98-9f6d-4bfa-abba-ee21c09f51df"). InnerVolumeSpecName "kube-api-access-f6kdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.182254 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-kube-api-access-ljkkl" (OuterVolumeSpecName: "kube-api-access-ljkkl") pod "8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" (UID: "8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c"). InnerVolumeSpecName "kube-api-access-ljkkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.223255 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" (UID: "8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.223761 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" (UID: "1d8a4b98-9f6d-4bfa-abba-ee21c09f51df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.267689 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-catalog-content\") pod \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.267782 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-operator-metrics\") pod \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.267844 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-utilities\") pod \"8f07095b-d178-4ef1-9269-b47e33d71ef3\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.267860 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-catalog-content\") pod \"8f07095b-d178-4ef1-9269-b47e33d71ef3\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.267880 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-utilities\") pod \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.267924 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-trusted-ca\") pod \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.267952 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk6sg\" (UniqueName: \"kubernetes.io/projected/69dee6c1-1c2f-4d49-8955-ca4226b4306e-kube-api-access-fk6sg\") pod \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\" (UID: \"69dee6c1-1c2f-4d49-8955-ca4226b4306e\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.267988 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v8h2\" (UniqueName: \"kubernetes.io/projected/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-kube-api-access-8v8h2\") pod \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\" (UID: \"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.268045 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prgw4\" (UniqueName: \"kubernetes.io/projected/8f07095b-d178-4ef1-9269-b47e33d71ef3-kube-api-access-prgw4\") pod \"8f07095b-d178-4ef1-9269-b47e33d71ef3\" (UID: \"8f07095b-d178-4ef1-9269-b47e33d71ef3\") " Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.268915 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-utilities" (OuterVolumeSpecName: "utilities") pod "8f07095b-d178-4ef1-9269-b47e33d71ef3" (UID: "8f07095b-d178-4ef1-9269-b47e33d71ef3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.269208 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.269225 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.269234 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljkkl\" (UniqueName: \"kubernetes.io/projected/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-kube-api-access-ljkkl\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.269270 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.269280 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.269289 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6kdp\" (UniqueName: \"kubernetes.io/projected/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df-kube-api-access-f6kdp\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.269298 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.269584 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "69dee6c1-1c2f-4d49-8955-ca4226b4306e" (UID: "69dee6c1-1c2f-4d49-8955-ca4226b4306e"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.270184 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-utilities" (OuterVolumeSpecName: "utilities") pod "5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" (UID: "5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.272369 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69dee6c1-1c2f-4d49-8955-ca4226b4306e-kube-api-access-fk6sg" (OuterVolumeSpecName: "kube-api-access-fk6sg") pod "69dee6c1-1c2f-4d49-8955-ca4226b4306e" (UID: "69dee6c1-1c2f-4d49-8955-ca4226b4306e"). InnerVolumeSpecName "kube-api-access-fk6sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.273210 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "69dee6c1-1c2f-4d49-8955-ca4226b4306e" (UID: "69dee6c1-1c2f-4d49-8955-ca4226b4306e"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.273356 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-kube-api-access-8v8h2" (OuterVolumeSpecName: "kube-api-access-8v8h2") pod "5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" (UID: "5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4"). InnerVolumeSpecName "kube-api-access-8v8h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.278730 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f07095b-d178-4ef1-9269-b47e33d71ef3-kube-api-access-prgw4" (OuterVolumeSpecName: "kube-api-access-prgw4") pod "8f07095b-d178-4ef1-9269-b47e33d71ef3" (UID: "8f07095b-d178-4ef1-9269-b47e33d71ef3"). InnerVolumeSpecName "kube-api-access-prgw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.293887 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" (UID: "5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.322844 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ksmlj"] Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.370205 4751 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.370237 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.370248 4751 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/69dee6c1-1c2f-4d49-8955-ca4226b4306e-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.370257 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk6sg\" (UniqueName: \"kubernetes.io/projected/69dee6c1-1c2f-4d49-8955-ca4226b4306e-kube-api-access-fk6sg\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.370267 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v8h2\" (UniqueName: \"kubernetes.io/projected/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-kube-api-access-8v8h2\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.370277 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prgw4\" (UniqueName: \"kubernetes.io/projected/8f07095b-d178-4ef1-9269-b47e33d71ef3-kube-api-access-prgw4\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.370288 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.429385 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f07095b-d178-4ef1-9269-b47e33d71ef3" (UID: "8f07095b-d178-4ef1-9269-b47e33d71ef3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.471041 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f07095b-d178-4ef1-9269-b47e33d71ef3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.871918 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqzbn" event={"ID":"8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c","Type":"ContainerDied","Data":"3588e02fe399b6cf2ef41a37dd4b437bb0fd052f6a59909d9ca73efa576c3923"} Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.871955 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqzbn" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.872013 4751 scope.go:117] "RemoveContainer" containerID="f64239b9f399ae12b962432d2a8d18f2c66c15c8e2cbdf468060f94c05b1103b" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.875877 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pc9vc" event={"ID":"5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4","Type":"ContainerDied","Data":"d9a1781f1af93f7d7bef98eff88870c22b226e91684879ca3247afc9d6a197a4"} Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.876014 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pc9vc" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.881728 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tlrgw" event={"ID":"8f07095b-d178-4ef1-9269-b47e33d71ef3","Type":"ContainerDied","Data":"55454a5f67103e7241a52b787a94b012b1a2f17dda2afcbab9c898cc061e5c08"} Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.881834 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tlrgw" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.884717 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" event={"ID":"9674dab9-cf73-46a0-8077-1ec770c4f73c","Type":"ContainerStarted","Data":"5e9bb949d0927fd65361b5f4a041a182b2b2888f0cf9ab62eb7dd3954b735716"} Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.884776 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" event={"ID":"9674dab9-cf73-46a0-8077-1ec770c4f73c","Type":"ContainerStarted","Data":"02445dc8c9a92b64fe036c2e73f68a05e14f7b628f3e6e5b306d67db54258e87"} Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.884973 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.887022 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" event={"ID":"69dee6c1-1c2f-4d49-8955-ca4226b4306e","Type":"ContainerDied","Data":"f38fceacb6bd0d734314c04b36462819587a88449c72583b762c2fe4773dd5a4"} Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.887131 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sph8v" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.890065 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.898507 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kc6w" event={"ID":"1d8a4b98-9f6d-4bfa-abba-ee21c09f51df","Type":"ContainerDied","Data":"040740a1624631b07ad2604d55932e813ad46029a4833aa4d0afe1e89736843b"} Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.898568 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8kc6w" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.902880 4751 scope.go:117] "RemoveContainer" containerID="17caa63336ce668a7f6c0d0b5bba1aa4ca99fd48874bb0a1946558ec6e9ab23a" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.916463 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ksmlj" podStartSLOduration=1.916441549 podStartE2EDuration="1.916441549s" podCreationTimestamp="2026-03-16 00:12:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:12:23.90681917 +0000 UTC m=+372.176151470" watchObservedRunningTime="2026-03-16 00:12:23.916441549 +0000 UTC m=+372.185773849" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.933029 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqzbn"] Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.949911 4751 scope.go:117] "RemoveContainer" containerID="2609e29cdc6919c3ca133323fb2abfbeca46e9dbdb36bac936940f849504b789" Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.953147 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mqzbn"] Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.986408 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8kc6w"] Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.993412 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8kc6w"] Mar 16 00:12:23 crc kubenswrapper[4751]: I0316 00:12:23.995497 4751 scope.go:117] "RemoveContainer" containerID="7932415c0e6d1902552b5f52c21327885c0b12df275ce4e58d0a0259b3a42414" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.001289 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sph8v"] Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.006852 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sph8v"] Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.013712 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tlrgw"] Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.018028 4751 scope.go:117] "RemoveContainer" containerID="3dd633f9cc74ab824e8193eefb7e5b4b077573c7e246929860b3cf19cdcf7a29" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.019737 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tlrgw"] Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.023724 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc9vc"] Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.026305 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pc9vc"] Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.033963 4751 scope.go:117] "RemoveContainer" containerID="16d6ec0a01195cb2eb46262c506491c9603b0788a1b67dd529ffaeb9995ec261" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.060414 4751 scope.go:117] "RemoveContainer" containerID="30ba8ff4e31de5142bcf76a3e58621e7063f410fb08e9cc5e4c48a96c58c9ac7" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.084651 4751 scope.go:117] "RemoveContainer" containerID="6342b3764cb725d4f3ae4e9c388a82ae6908a92cf22be82cb7424a9e7686310b" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.097739 4751 scope.go:117] "RemoveContainer" containerID="eea43c3512fed11d289e063414da7264c2d03b1b3f04b97c9d8cd6cc975bb6f3" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.109057 4751 scope.go:117] "RemoveContainer" containerID="9ba145bd0d82918f5fcf95cb700001241529e0f4b695b8cead72e61eef415a21" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.121675 4751 scope.go:117] "RemoveContainer" containerID="7c86bd0dab7affbe02238e2fc8a5765d49551ed62988d987b925aa2e3f691f7a" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.135550 4751 scope.go:117] "RemoveContainer" containerID="91b3e81da9a9f336743af63c56b2ace29ce282f1c0ed420dac4656407b512f51" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.155749 4751 scope.go:117] "RemoveContainer" containerID="65ea2db385173b1f397b1d71582271814011a09be369aa8bf4d849ab3b8be80e" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.465178 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" path="/var/lib/kubelet/pods/1d8a4b98-9f6d-4bfa-abba-ee21c09f51df/volumes" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.465834 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" path="/var/lib/kubelet/pods/5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4/volumes" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.466500 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" path="/var/lib/kubelet/pods/69dee6c1-1c2f-4d49-8955-ca4226b4306e/volumes" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.466935 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" path="/var/lib/kubelet/pods/8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c/volumes" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.467493 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" path="/var/lib/kubelet/pods/8f07095b-d178-4ef1-9269-b47e33d71ef3/volumes" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.811026 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mp6q2"] Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812494 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812505 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812513 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerName="extract-utilities" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812519 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerName="extract-utilities" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812528 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerName="extract-content" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812535 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerName="extract-content" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812542 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812548 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812558 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerName="extract-utilities" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812565 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerName="extract-utilities" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812574 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerName="marketplace-operator" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812580 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerName="marketplace-operator" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812586 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812592 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812599 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerName="extract-utilities" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812605 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerName="extract-utilities" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812613 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerName="extract-content" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812619 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerName="extract-content" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812629 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812635 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812643 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerName="extract-content" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812649 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerName="extract-content" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812656 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerName="extract-content" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812661 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerName="extract-content" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812671 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerName="extract-utilities" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812676 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerName="extract-utilities" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812768 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ab4fc9b-0233-44ec-8388-38dbd4eb6d9c" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812781 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerName="marketplace-operator" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812788 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b19ec9b-b56e-4e9f-a411-b5ca8013f2e4" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812795 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f07095b-d178-4ef1-9269-b47e33d71ef3" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812803 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8a4b98-9f6d-4bfa-abba-ee21c09f51df" containerName="registry-server" Mar 16 00:12:24 crc kubenswrapper[4751]: E0316 00:12:24.812875 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerName="marketplace-operator" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812882 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerName="marketplace-operator" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.812960 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="69dee6c1-1c2f-4d49-8955-ca4226b4306e" containerName="marketplace-operator" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.813447 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.817255 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.851771 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mp6q2"] Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.887550 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-catalog-content\") pod \"redhat-marketplace-mp6q2\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.887661 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-utilities\") pod \"redhat-marketplace-mp6q2\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.887743 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5wrv\" (UniqueName: \"kubernetes.io/projected/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-kube-api-access-r5wrv\") pod \"redhat-marketplace-mp6q2\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.989379 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-catalog-content\") pod \"redhat-marketplace-mp6q2\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.989506 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-utilities\") pod \"redhat-marketplace-mp6q2\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.990275 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-catalog-content\") pod \"redhat-marketplace-mp6q2\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.990330 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5wrv\" (UniqueName: \"kubernetes.io/projected/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-kube-api-access-r5wrv\") pod \"redhat-marketplace-mp6q2\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:24 crc kubenswrapper[4751]: I0316 00:12:24.990376 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-utilities\") pod \"redhat-marketplace-mp6q2\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.013658 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-57pn4"] Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.013995 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5wrv\" (UniqueName: \"kubernetes.io/projected/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-kube-api-access-r5wrv\") pod \"redhat-marketplace-mp6q2\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.015627 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.018568 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.028624 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-57pn4"] Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.091138 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-utilities\") pod \"certified-operators-57pn4\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.091204 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27k94\" (UniqueName: \"kubernetes.io/projected/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-kube-api-access-27k94\") pod \"certified-operators-57pn4\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.091403 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-catalog-content\") pod \"certified-operators-57pn4\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.151829 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.195324 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-utilities\") pod \"certified-operators-57pn4\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.195452 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27k94\" (UniqueName: \"kubernetes.io/projected/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-kube-api-access-27k94\") pod \"certified-operators-57pn4\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.195575 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-catalog-content\") pod \"certified-operators-57pn4\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.196271 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-utilities\") pod \"certified-operators-57pn4\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.196638 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-catalog-content\") pod \"certified-operators-57pn4\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.230964 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27k94\" (UniqueName: \"kubernetes.io/projected/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-kube-api-access-27k94\") pod \"certified-operators-57pn4\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:25 crc kubenswrapper[4751]: I0316 00:12:25.375485 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:26 crc kubenswrapper[4751]: I0316 00:12:25.564550 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-57pn4"] Mar 16 00:12:26 crc kubenswrapper[4751]: W0316 00:12:25.565484 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76c15cfb_0b0f_4f9f_94eb_22b0d736f175.slice/crio-dad2568a386f3923d70b0f03f2775b7409dcf80746d7e5e9bc0d874aded1a2a9 WatchSource:0}: Error finding container dad2568a386f3923d70b0f03f2775b7409dcf80746d7e5e9bc0d874aded1a2a9: Status 404 returned error can't find the container with id dad2568a386f3923d70b0f03f2775b7409dcf80746d7e5e9bc0d874aded1a2a9 Mar 16 00:12:26 crc kubenswrapper[4751]: I0316 00:12:25.592605 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mp6q2"] Mar 16 00:12:26 crc kubenswrapper[4751]: W0316 00:12:25.600900 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bd04a02_93f5_4cd5_88c9_bccc3cae359f.slice/crio-b0f9fae330d209d0b9963d885a2feae7f5e9ea500fea40a06032e74af6fc2853 WatchSource:0}: Error finding container b0f9fae330d209d0b9963d885a2feae7f5e9ea500fea40a06032e74af6fc2853: Status 404 returned error can't find the container with id b0f9fae330d209d0b9963d885a2feae7f5e9ea500fea40a06032e74af6fc2853 Mar 16 00:12:26 crc kubenswrapper[4751]: I0316 00:12:25.918317 4751 generic.go:334] "Generic (PLEG): container finished" podID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerID="e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b" exitCode=0 Mar 16 00:12:26 crc kubenswrapper[4751]: I0316 00:12:25.918387 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57pn4" event={"ID":"76c15cfb-0b0f-4f9f-94eb-22b0d736f175","Type":"ContainerDied","Data":"e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b"} Mar 16 00:12:26 crc kubenswrapper[4751]: I0316 00:12:25.918415 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57pn4" event={"ID":"76c15cfb-0b0f-4f9f-94eb-22b0d736f175","Type":"ContainerStarted","Data":"dad2568a386f3923d70b0f03f2775b7409dcf80746d7e5e9bc0d874aded1a2a9"} Mar 16 00:12:26 crc kubenswrapper[4751]: I0316 00:12:25.922646 4751 generic.go:334] "Generic (PLEG): container finished" podID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerID="cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca" exitCode=0 Mar 16 00:12:26 crc kubenswrapper[4751]: I0316 00:12:25.922741 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mp6q2" event={"ID":"3bd04a02-93f5-4cd5-88c9-bccc3cae359f","Type":"ContainerDied","Data":"cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca"} Mar 16 00:12:26 crc kubenswrapper[4751]: I0316 00:12:25.922800 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mp6q2" event={"ID":"3bd04a02-93f5-4cd5-88c9-bccc3cae359f","Type":"ContainerStarted","Data":"b0f9fae330d209d0b9963d885a2feae7f5e9ea500fea40a06032e74af6fc2853"} Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.214404 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j8n4q"] Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.218707 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.219921 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j8n4q"] Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.221244 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.324503 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1568ce0c-e5c5-4fa1-a95b-470216db436f-utilities\") pod \"redhat-operators-j8n4q\" (UID: \"1568ce0c-e5c5-4fa1-a95b-470216db436f\") " pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.324629 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzlcj\" (UniqueName: \"kubernetes.io/projected/1568ce0c-e5c5-4fa1-a95b-470216db436f-kube-api-access-lzlcj\") pod \"redhat-operators-j8n4q\" (UID: \"1568ce0c-e5c5-4fa1-a95b-470216db436f\") " pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.324713 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1568ce0c-e5c5-4fa1-a95b-470216db436f-catalog-content\") pod \"redhat-operators-j8n4q\" (UID: \"1568ce0c-e5c5-4fa1-a95b-470216db436f\") " pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.410756 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nnkxp"] Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.411743 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.414877 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.429384 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nnkxp"] Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.430095 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzlcj\" (UniqueName: \"kubernetes.io/projected/1568ce0c-e5c5-4fa1-a95b-470216db436f-kube-api-access-lzlcj\") pod \"redhat-operators-j8n4q\" (UID: \"1568ce0c-e5c5-4fa1-a95b-470216db436f\") " pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.430197 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1568ce0c-e5c5-4fa1-a95b-470216db436f-catalog-content\") pod \"redhat-operators-j8n4q\" (UID: \"1568ce0c-e5c5-4fa1-a95b-470216db436f\") " pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.430305 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1568ce0c-e5c5-4fa1-a95b-470216db436f-utilities\") pod \"redhat-operators-j8n4q\" (UID: \"1568ce0c-e5c5-4fa1-a95b-470216db436f\") " pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.431014 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1568ce0c-e5c5-4fa1-a95b-470216db436f-utilities\") pod \"redhat-operators-j8n4q\" (UID: \"1568ce0c-e5c5-4fa1-a95b-470216db436f\") " pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.431812 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1568ce0c-e5c5-4fa1-a95b-470216db436f-catalog-content\") pod \"redhat-operators-j8n4q\" (UID: \"1568ce0c-e5c5-4fa1-a95b-470216db436f\") " pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.486235 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzlcj\" (UniqueName: \"kubernetes.io/projected/1568ce0c-e5c5-4fa1-a95b-470216db436f-kube-api-access-lzlcj\") pod \"redhat-operators-j8n4q\" (UID: \"1568ce0c-e5c5-4fa1-a95b-470216db436f\") " pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.531097 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw695\" (UniqueName: \"kubernetes.io/projected/647a9f72-750d-4c57-8684-4c841604a40e-kube-api-access-jw695\") pod \"community-operators-nnkxp\" (UID: \"647a9f72-750d-4c57-8684-4c841604a40e\") " pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.531352 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/647a9f72-750d-4c57-8684-4c841604a40e-catalog-content\") pod \"community-operators-nnkxp\" (UID: \"647a9f72-750d-4c57-8684-4c841604a40e\") " pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.531443 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/647a9f72-750d-4c57-8684-4c841604a40e-utilities\") pod \"community-operators-nnkxp\" (UID: \"647a9f72-750d-4c57-8684-4c841604a40e\") " pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.544443 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.641059 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/647a9f72-750d-4c57-8684-4c841604a40e-utilities\") pod \"community-operators-nnkxp\" (UID: \"647a9f72-750d-4c57-8684-4c841604a40e\") " pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.641509 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw695\" (UniqueName: \"kubernetes.io/projected/647a9f72-750d-4c57-8684-4c841604a40e-kube-api-access-jw695\") pod \"community-operators-nnkxp\" (UID: \"647a9f72-750d-4c57-8684-4c841604a40e\") " pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.641644 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/647a9f72-750d-4c57-8684-4c841604a40e-catalog-content\") pod \"community-operators-nnkxp\" (UID: \"647a9f72-750d-4c57-8684-4c841604a40e\") " pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.642738 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/647a9f72-750d-4c57-8684-4c841604a40e-utilities\") pod \"community-operators-nnkxp\" (UID: \"647a9f72-750d-4c57-8684-4c841604a40e\") " pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.642865 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/647a9f72-750d-4c57-8684-4c841604a40e-catalog-content\") pod \"community-operators-nnkxp\" (UID: \"647a9f72-750d-4c57-8684-4c841604a40e\") " pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.673041 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw695\" (UniqueName: \"kubernetes.io/projected/647a9f72-750d-4c57-8684-4c841604a40e-kube-api-access-jw695\") pod \"community-operators-nnkxp\" (UID: \"647a9f72-750d-4c57-8684-4c841604a40e\") " pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.784929 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.934926 4751 generic.go:334] "Generic (PLEG): container finished" podID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerID="32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5" exitCode=0 Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.935417 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mp6q2" event={"ID":"3bd04a02-93f5-4cd5-88c9-bccc3cae359f","Type":"ContainerDied","Data":"32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5"} Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.939081 4751 generic.go:334] "Generic (PLEG): container finished" podID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerID="11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c" exitCode=0 Mar 16 00:12:27 crc kubenswrapper[4751]: I0316 00:12:27.939135 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57pn4" event={"ID":"76c15cfb-0b0f-4f9f-94eb-22b0d736f175","Type":"ContainerDied","Data":"11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c"} Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.006599 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j8n4q"] Mar 16 00:12:28 crc kubenswrapper[4751]: W0316 00:12:28.008945 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1568ce0c_e5c5_4fa1_a95b_470216db436f.slice/crio-8b7c5aff0e87013ada5a3de6e92e3c645cb9e7feda172d82f5b77d8c58881a25 WatchSource:0}: Error finding container 8b7c5aff0e87013ada5a3de6e92e3c645cb9e7feda172d82f5b77d8c58881a25: Status 404 returned error can't find the container with id 8b7c5aff0e87013ada5a3de6e92e3c645cb9e7feda172d82f5b77d8c58881a25 Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.168717 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nnkxp"] Mar 16 00:12:28 crc kubenswrapper[4751]: W0316 00:12:28.172735 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod647a9f72_750d_4c57_8684_4c841604a40e.slice/crio-a880a84f814239039036f32c4dccc40416578429997a0ae9bddb6d9f50903d79 WatchSource:0}: Error finding container a880a84f814239039036f32c4dccc40416578429997a0ae9bddb6d9f50903d79: Status 404 returned error can't find the container with id a880a84f814239039036f32c4dccc40416578429997a0ae9bddb6d9f50903d79 Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.974069 4751 generic.go:334] "Generic (PLEG): container finished" podID="1568ce0c-e5c5-4fa1-a95b-470216db436f" containerID="daa54d2cec5ca2099603f2abe9308a5f78b02df784f7cda4d36cce4a90be10b7" exitCode=0 Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.974160 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8n4q" event={"ID":"1568ce0c-e5c5-4fa1-a95b-470216db436f","Type":"ContainerDied","Data":"daa54d2cec5ca2099603f2abe9308a5f78b02df784f7cda4d36cce4a90be10b7"} Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.974542 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8n4q" event={"ID":"1568ce0c-e5c5-4fa1-a95b-470216db436f","Type":"ContainerStarted","Data":"8b7c5aff0e87013ada5a3de6e92e3c645cb9e7feda172d82f5b77d8c58881a25"} Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.977549 4751 generic.go:334] "Generic (PLEG): container finished" podID="647a9f72-750d-4c57-8684-4c841604a40e" containerID="03dc522d34737ccd55c7f694b3d8fb81d430d719bd23567f27e6e3cbf3db2b65" exitCode=0 Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.977672 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nnkxp" event={"ID":"647a9f72-750d-4c57-8684-4c841604a40e","Type":"ContainerDied","Data":"03dc522d34737ccd55c7f694b3d8fb81d430d719bd23567f27e6e3cbf3db2b65"} Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.977710 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nnkxp" event={"ID":"647a9f72-750d-4c57-8684-4c841604a40e","Type":"ContainerStarted","Data":"a880a84f814239039036f32c4dccc40416578429997a0ae9bddb6d9f50903d79"} Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.980715 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57pn4" event={"ID":"76c15cfb-0b0f-4f9f-94eb-22b0d736f175","Type":"ContainerStarted","Data":"87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1"} Mar 16 00:12:28 crc kubenswrapper[4751]: I0316 00:12:28.983773 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mp6q2" event={"ID":"3bd04a02-93f5-4cd5-88c9-bccc3cae359f","Type":"ContainerStarted","Data":"85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66"} Mar 16 00:12:29 crc kubenswrapper[4751]: I0316 00:12:29.024208 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-57pn4" podStartSLOduration=2.5417956569999998 podStartE2EDuration="5.024189568s" podCreationTimestamp="2026-03-16 00:12:24 +0000 UTC" firstStartedPulling="2026-03-16 00:12:25.919962393 +0000 UTC m=+374.189294683" lastFinishedPulling="2026-03-16 00:12:28.402356314 +0000 UTC m=+376.671688594" observedRunningTime="2026-03-16 00:12:29.021436851 +0000 UTC m=+377.290769191" watchObservedRunningTime="2026-03-16 00:12:29.024189568 +0000 UTC m=+377.293521878" Mar 16 00:12:29 crc kubenswrapper[4751]: I0316 00:12:29.051905 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mp6q2" podStartSLOduration=2.291908154 podStartE2EDuration="5.051885853s" podCreationTimestamp="2026-03-16 00:12:24 +0000 UTC" firstStartedPulling="2026-03-16 00:12:25.923993335 +0000 UTC m=+374.193325625" lastFinishedPulling="2026-03-16 00:12:28.683971024 +0000 UTC m=+376.953303324" observedRunningTime="2026-03-16 00:12:29.048040738 +0000 UTC m=+377.317373028" watchObservedRunningTime="2026-03-16 00:12:29.051885853 +0000 UTC m=+377.321218153" Mar 16 00:12:29 crc kubenswrapper[4751]: I0316 00:12:29.992190 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8n4q" event={"ID":"1568ce0c-e5c5-4fa1-a95b-470216db436f","Type":"ContainerStarted","Data":"833d1e9fb799bbd81daa2604d899d8e8ed0d6d0e4a19705d0c129771dbab1b7f"} Mar 16 00:12:29 crc kubenswrapper[4751]: I0316 00:12:29.994032 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nnkxp" event={"ID":"647a9f72-750d-4c57-8684-4c841604a40e","Type":"ContainerStarted","Data":"eeaf1e03f8790977c08120502a802e8f507f6f0f38d552df51c1c6f4f131bd64"} Mar 16 00:12:31 crc kubenswrapper[4751]: I0316 00:12:31.001966 4751 generic.go:334] "Generic (PLEG): container finished" podID="647a9f72-750d-4c57-8684-4c841604a40e" containerID="eeaf1e03f8790977c08120502a802e8f507f6f0f38d552df51c1c6f4f131bd64" exitCode=0 Mar 16 00:12:31 crc kubenswrapper[4751]: I0316 00:12:31.002046 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nnkxp" event={"ID":"647a9f72-750d-4c57-8684-4c841604a40e","Type":"ContainerDied","Data":"eeaf1e03f8790977c08120502a802e8f507f6f0f38d552df51c1c6f4f131bd64"} Mar 16 00:12:31 crc kubenswrapper[4751]: I0316 00:12:31.005163 4751 generic.go:334] "Generic (PLEG): container finished" podID="1568ce0c-e5c5-4fa1-a95b-470216db436f" containerID="833d1e9fb799bbd81daa2604d899d8e8ed0d6d0e4a19705d0c129771dbab1b7f" exitCode=0 Mar 16 00:12:31 crc kubenswrapper[4751]: I0316 00:12:31.005193 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8n4q" event={"ID":"1568ce0c-e5c5-4fa1-a95b-470216db436f","Type":"ContainerDied","Data":"833d1e9fb799bbd81daa2604d899d8e8ed0d6d0e4a19705d0c129771dbab1b7f"} Mar 16 00:12:33 crc kubenswrapper[4751]: I0316 00:12:33.021805 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j8n4q" event={"ID":"1568ce0c-e5c5-4fa1-a95b-470216db436f","Type":"ContainerStarted","Data":"9ab6a4573ab3a1e23cb8df170cfc39c882f4b1bb1a84d92de807a91f93c26795"} Mar 16 00:12:33 crc kubenswrapper[4751]: I0316 00:12:33.024187 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nnkxp" event={"ID":"647a9f72-750d-4c57-8684-4c841604a40e","Type":"ContainerStarted","Data":"f758044467817f887e64c9dfa8f995509f97a40437fc6d722a52873cca55e89b"} Mar 16 00:12:33 crc kubenswrapper[4751]: I0316 00:12:33.046696 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j8n4q" podStartSLOduration=2.405829771 podStartE2EDuration="6.0466729s" podCreationTimestamp="2026-03-16 00:12:27 +0000 UTC" firstStartedPulling="2026-03-16 00:12:28.975957869 +0000 UTC m=+377.245290179" lastFinishedPulling="2026-03-16 00:12:32.616801008 +0000 UTC m=+380.886133308" observedRunningTime="2026-03-16 00:12:33.04440769 +0000 UTC m=+381.313739990" watchObservedRunningTime="2026-03-16 00:12:33.0466729 +0000 UTC m=+381.316005210" Mar 16 00:12:33 crc kubenswrapper[4751]: I0316 00:12:33.075596 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nnkxp" podStartSLOduration=2.37117225 podStartE2EDuration="6.075575498s" podCreationTimestamp="2026-03-16 00:12:27 +0000 UTC" firstStartedPulling="2026-03-16 00:12:28.978628383 +0000 UTC m=+377.247960673" lastFinishedPulling="2026-03-16 00:12:32.683031591 +0000 UTC m=+380.952363921" observedRunningTime="2026-03-16 00:12:33.072014922 +0000 UTC m=+381.341347242" watchObservedRunningTime="2026-03-16 00:12:33.075575498 +0000 UTC m=+381.344907808" Mar 16 00:12:35 crc kubenswrapper[4751]: I0316 00:12:35.152240 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:35 crc kubenswrapper[4751]: I0316 00:12:35.152630 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:35 crc kubenswrapper[4751]: I0316 00:12:35.247116 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:35 crc kubenswrapper[4751]: I0316 00:12:35.376449 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:35 crc kubenswrapper[4751]: I0316 00:12:35.376515 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:35 crc kubenswrapper[4751]: I0316 00:12:35.422990 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:36 crc kubenswrapper[4751]: I0316 00:12:36.103215 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:12:36 crc kubenswrapper[4751]: I0316 00:12:36.104165 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:12:37 crc kubenswrapper[4751]: I0316 00:12:37.545096 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:37 crc kubenswrapper[4751]: I0316 00:12:37.545166 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:37 crc kubenswrapper[4751]: I0316 00:12:37.786039 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:37 crc kubenswrapper[4751]: I0316 00:12:37.786422 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:37 crc kubenswrapper[4751]: I0316 00:12:37.823820 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:38 crc kubenswrapper[4751]: I0316 00:12:38.096733 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nnkxp" Mar 16 00:12:38 crc kubenswrapper[4751]: I0316 00:12:38.590763 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j8n4q" podUID="1568ce0c-e5c5-4fa1-a95b-470216db436f" containerName="registry-server" probeResult="failure" output=< Mar 16 00:12:38 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Mar 16 00:12:38 crc kubenswrapper[4751]: > Mar 16 00:12:47 crc kubenswrapper[4751]: I0316 00:12:47.604660 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:12:47 crc kubenswrapper[4751]: I0316 00:12:47.659238 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j8n4q" Mar 16 00:13:57 crc kubenswrapper[4751]: I0316 00:13:57.853920 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:13:57 crc kubenswrapper[4751]: I0316 00:13:57.854464 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.152589 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560334-5w59w"] Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.154234 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560334-5w59w" Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.160798 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.161302 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.161690 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.187317 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560334-5w59w"] Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.228153 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vff5q\" (UniqueName: \"kubernetes.io/projected/6aa4a917-7cee-46f0-ad11-1fcdeacb0469-kube-api-access-vff5q\") pod \"auto-csr-approver-29560334-5w59w\" (UID: \"6aa4a917-7cee-46f0-ad11-1fcdeacb0469\") " pod="openshift-infra/auto-csr-approver-29560334-5w59w" Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.329239 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vff5q\" (UniqueName: \"kubernetes.io/projected/6aa4a917-7cee-46f0-ad11-1fcdeacb0469-kube-api-access-vff5q\") pod \"auto-csr-approver-29560334-5w59w\" (UID: \"6aa4a917-7cee-46f0-ad11-1fcdeacb0469\") " pod="openshift-infra/auto-csr-approver-29560334-5w59w" Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.365170 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vff5q\" (UniqueName: \"kubernetes.io/projected/6aa4a917-7cee-46f0-ad11-1fcdeacb0469-kube-api-access-vff5q\") pod \"auto-csr-approver-29560334-5w59w\" (UID: \"6aa4a917-7cee-46f0-ad11-1fcdeacb0469\") " pod="openshift-infra/auto-csr-approver-29560334-5w59w" Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.492331 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560334-5w59w" Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.748010 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560334-5w59w"] Mar 16 00:14:00 crc kubenswrapper[4751]: I0316 00:14:00.756646 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 16 00:14:01 crc kubenswrapper[4751]: I0316 00:14:01.631705 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560334-5w59w" event={"ID":"6aa4a917-7cee-46f0-ad11-1fcdeacb0469","Type":"ContainerStarted","Data":"7d6a58e6cf170b81ca04fea35489aaf6e90dbc2774d5102827e9a3b54b7d1c39"} Mar 16 00:14:02 crc kubenswrapper[4751]: I0316 00:14:02.643989 4751 generic.go:334] "Generic (PLEG): container finished" podID="6aa4a917-7cee-46f0-ad11-1fcdeacb0469" containerID="2bff3cf5eda1ea9b616976a38f61facd57dfd144f2a9db25cc1700d1a1d9bd03" exitCode=0 Mar 16 00:14:02 crc kubenswrapper[4751]: I0316 00:14:02.644058 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560334-5w59w" event={"ID":"6aa4a917-7cee-46f0-ad11-1fcdeacb0469","Type":"ContainerDied","Data":"2bff3cf5eda1ea9b616976a38f61facd57dfd144f2a9db25cc1700d1a1d9bd03"} Mar 16 00:14:03 crc kubenswrapper[4751]: I0316 00:14:03.919144 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560334-5w59w" Mar 16 00:14:03 crc kubenswrapper[4751]: I0316 00:14:03.986283 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vff5q\" (UniqueName: \"kubernetes.io/projected/6aa4a917-7cee-46f0-ad11-1fcdeacb0469-kube-api-access-vff5q\") pod \"6aa4a917-7cee-46f0-ad11-1fcdeacb0469\" (UID: \"6aa4a917-7cee-46f0-ad11-1fcdeacb0469\") " Mar 16 00:14:03 crc kubenswrapper[4751]: I0316 00:14:03.991208 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa4a917-7cee-46f0-ad11-1fcdeacb0469-kube-api-access-vff5q" (OuterVolumeSpecName: "kube-api-access-vff5q") pod "6aa4a917-7cee-46f0-ad11-1fcdeacb0469" (UID: "6aa4a917-7cee-46f0-ad11-1fcdeacb0469"). InnerVolumeSpecName "kube-api-access-vff5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:14:04 crc kubenswrapper[4751]: I0316 00:14:04.087691 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vff5q\" (UniqueName: \"kubernetes.io/projected/6aa4a917-7cee-46f0-ad11-1fcdeacb0469-kube-api-access-vff5q\") on node \"crc\" DevicePath \"\"" Mar 16 00:14:04 crc kubenswrapper[4751]: I0316 00:14:04.678442 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560334-5w59w" event={"ID":"6aa4a917-7cee-46f0-ad11-1fcdeacb0469","Type":"ContainerDied","Data":"7d6a58e6cf170b81ca04fea35489aaf6e90dbc2774d5102827e9a3b54b7d1c39"} Mar 16 00:14:04 crc kubenswrapper[4751]: I0316 00:14:04.678747 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d6a58e6cf170b81ca04fea35489aaf6e90dbc2774d5102827e9a3b54b7d1c39" Mar 16 00:14:04 crc kubenswrapper[4751]: I0316 00:14:04.678530 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560334-5w59w" Mar 16 00:14:04 crc kubenswrapper[4751]: I0316 00:14:04.985745 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29560328-lt56z"] Mar 16 00:14:04 crc kubenswrapper[4751]: I0316 00:14:04.987644 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29560328-lt56z"] Mar 16 00:14:06 crc kubenswrapper[4751]: I0316 00:14:06.469680 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05" path="/var/lib/kubelet/pods/c4dc7d4c-9528-4cf4-a8f9-59648c8fbf05/volumes" Mar 16 00:14:27 crc kubenswrapper[4751]: I0316 00:14:27.854633 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:14:27 crc kubenswrapper[4751]: I0316 00:14:27.855392 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:14:57 crc kubenswrapper[4751]: I0316 00:14:57.854527 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:14:57 crc kubenswrapper[4751]: I0316 00:14:57.855268 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:14:57 crc kubenswrapper[4751]: I0316 00:14:57.855329 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:14:57 crc kubenswrapper[4751]: I0316 00:14:57.856003 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"36270f1e57b171978cc55e8e0d2e8251ea29ff34e7dce546d8c65c5b8b799266"} pod="openshift-machine-config-operator/machine-config-daemon-6fccz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 16 00:14:57 crc kubenswrapper[4751]: I0316 00:14:57.856086 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" containerID="cri-o://36270f1e57b171978cc55e8e0d2e8251ea29ff34e7dce546d8c65c5b8b799266" gracePeriod=600 Mar 16 00:14:58 crc kubenswrapper[4751]: I0316 00:14:58.066444 4751 generic.go:334] "Generic (PLEG): container finished" podID="25429719-d18d-4909-8455-701d09423c36" containerID="36270f1e57b171978cc55e8e0d2e8251ea29ff34e7dce546d8c65c5b8b799266" exitCode=0 Mar 16 00:14:58 crc kubenswrapper[4751]: I0316 00:14:58.066509 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerDied","Data":"36270f1e57b171978cc55e8e0d2e8251ea29ff34e7dce546d8c65c5b8b799266"} Mar 16 00:14:58 crc kubenswrapper[4751]: I0316 00:14:58.066564 4751 scope.go:117] "RemoveContainer" containerID="eb29dfbd6c43f45308629c345a9e5bfbd8b4e1c106c34f4c658c32496548540a" Mar 16 00:14:59 crc kubenswrapper[4751]: I0316 00:14:59.078370 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerStarted","Data":"a45820bd88f0ba2e569a216d14a8ef916b60532a9717b18faa5c4a39cc910750"} Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.207649 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn"] Mar 16 00:15:00 crc kubenswrapper[4751]: E0316 00:15:00.208049 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa4a917-7cee-46f0-ad11-1fcdeacb0469" containerName="oc" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.208062 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa4a917-7cee-46f0-ad11-1fcdeacb0469" containerName="oc" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.208284 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa4a917-7cee-46f0-ad11-1fcdeacb0469" containerName="oc" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.208767 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.223815 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.224013 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.227761 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn"] Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.382453 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-secret-volume\") pod \"collect-profiles-29560335-rw7dn\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.382550 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-config-volume\") pod \"collect-profiles-29560335-rw7dn\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.382739 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4wsw\" (UniqueName: \"kubernetes.io/projected/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-kube-api-access-t4wsw\") pod \"collect-profiles-29560335-rw7dn\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.485087 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4wsw\" (UniqueName: \"kubernetes.io/projected/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-kube-api-access-t4wsw\") pod \"collect-profiles-29560335-rw7dn\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.485250 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-secret-volume\") pod \"collect-profiles-29560335-rw7dn\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.485324 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-config-volume\") pod \"collect-profiles-29560335-rw7dn\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.486873 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-config-volume\") pod \"collect-profiles-29560335-rw7dn\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.501271 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-secret-volume\") pod \"collect-profiles-29560335-rw7dn\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.517006 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4wsw\" (UniqueName: \"kubernetes.io/projected/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-kube-api-access-t4wsw\") pod \"collect-profiles-29560335-rw7dn\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.528368 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:00 crc kubenswrapper[4751]: I0316 00:15:00.785394 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn"] Mar 16 00:15:01 crc kubenswrapper[4751]: I0316 00:15:01.095046 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" event={"ID":"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a","Type":"ContainerStarted","Data":"7345cfc0b7a877fd8cab2d60475f1020b992091a203607cd337becba517b7332"} Mar 16 00:15:01 crc kubenswrapper[4751]: I0316 00:15:01.095455 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" event={"ID":"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a","Type":"ContainerStarted","Data":"d55df8b9e09af81dcb1d6f39da5a7436560e7f6aa9a29af160a6c099f55f1acb"} Mar 16 00:15:02 crc kubenswrapper[4751]: I0316 00:15:02.103592 4751 generic.go:334] "Generic (PLEG): container finished" podID="90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a" containerID="7345cfc0b7a877fd8cab2d60475f1020b992091a203607cd337becba517b7332" exitCode=0 Mar 16 00:15:02 crc kubenswrapper[4751]: I0316 00:15:02.103763 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" event={"ID":"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a","Type":"ContainerDied","Data":"7345cfc0b7a877fd8cab2d60475f1020b992091a203607cd337becba517b7332"} Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.386614 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.426632 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4wsw\" (UniqueName: \"kubernetes.io/projected/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-kube-api-access-t4wsw\") pod \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.426712 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-config-volume\") pod \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.426767 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-secret-volume\") pod \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\" (UID: \"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a\") " Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.427652 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-config-volume" (OuterVolumeSpecName: "config-volume") pod "90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a" (UID: "90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.433065 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-kube-api-access-t4wsw" (OuterVolumeSpecName: "kube-api-access-t4wsw") pod "90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a" (UID: "90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a"). InnerVolumeSpecName "kube-api-access-t4wsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.436293 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a" (UID: "90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.528194 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.528245 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 16 00:15:03 crc kubenswrapper[4751]: I0316 00:15:03.528263 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4wsw\" (UniqueName: \"kubernetes.io/projected/90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a-kube-api-access-t4wsw\") on node \"crc\" DevicePath \"\"" Mar 16 00:15:04 crc kubenswrapper[4751]: I0316 00:15:04.118706 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" event={"ID":"90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a","Type":"ContainerDied","Data":"d55df8b9e09af81dcb1d6f39da5a7436560e7f6aa9a29af160a6c099f55f1acb"} Mar 16 00:15:04 crc kubenswrapper[4751]: I0316 00:15:04.118768 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560335-rw7dn" Mar 16 00:15:04 crc kubenswrapper[4751]: I0316 00:15:04.118773 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d55df8b9e09af81dcb1d6f39da5a7436560e7f6aa9a29af160a6c099f55f1acb" Mar 16 00:15:12 crc kubenswrapper[4751]: I0316 00:15:12.823350 4751 scope.go:117] "RemoveContainer" containerID="88d600e71538bb890b63c94cdfa3945488e931f89777334bdbf03d3f8e874009" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.156687 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560336-pvk5d"] Mar 16 00:16:00 crc kubenswrapper[4751]: E0316 00:16:00.157828 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a" containerName="collect-profiles" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.157853 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a" containerName="collect-profiles" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.158038 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="90a80ac1-4bde-4deb-9f2a-4b5b6ebd5a0a" containerName="collect-profiles" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.158877 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560336-pvk5d" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.162022 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.162836 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.162834 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.169589 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560336-pvk5d"] Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.202672 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxh57\" (UniqueName: \"kubernetes.io/projected/d9713fca-2f31-4266-809a-76c2085543ae-kube-api-access-rxh57\") pod \"auto-csr-approver-29560336-pvk5d\" (UID: \"d9713fca-2f31-4266-809a-76c2085543ae\") " pod="openshift-infra/auto-csr-approver-29560336-pvk5d" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.304393 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxh57\" (UniqueName: \"kubernetes.io/projected/d9713fca-2f31-4266-809a-76c2085543ae-kube-api-access-rxh57\") pod \"auto-csr-approver-29560336-pvk5d\" (UID: \"d9713fca-2f31-4266-809a-76c2085543ae\") " pod="openshift-infra/auto-csr-approver-29560336-pvk5d" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.338781 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxh57\" (UniqueName: \"kubernetes.io/projected/d9713fca-2f31-4266-809a-76c2085543ae-kube-api-access-rxh57\") pod \"auto-csr-approver-29560336-pvk5d\" (UID: \"d9713fca-2f31-4266-809a-76c2085543ae\") " pod="openshift-infra/auto-csr-approver-29560336-pvk5d" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.493354 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560336-pvk5d" Mar 16 00:16:00 crc kubenswrapper[4751]: I0316 00:16:00.748991 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560336-pvk5d"] Mar 16 00:16:01 crc kubenswrapper[4751]: I0316 00:16:01.524097 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560336-pvk5d" event={"ID":"d9713fca-2f31-4266-809a-76c2085543ae","Type":"ContainerStarted","Data":"373fe23b093afb2aa9921040f5da918e5bbb024cff1761b15af76b793f48cf26"} Mar 16 00:16:02 crc kubenswrapper[4751]: I0316 00:16:02.530807 4751 generic.go:334] "Generic (PLEG): container finished" podID="d9713fca-2f31-4266-809a-76c2085543ae" containerID="fcec6157b1ec83f0230be70b6805d656ab0b1106c392804f4eb26cffe0f43df3" exitCode=0 Mar 16 00:16:02 crc kubenswrapper[4751]: I0316 00:16:02.530898 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560336-pvk5d" event={"ID":"d9713fca-2f31-4266-809a-76c2085543ae","Type":"ContainerDied","Data":"fcec6157b1ec83f0230be70b6805d656ab0b1106c392804f4eb26cffe0f43df3"} Mar 16 00:16:03 crc kubenswrapper[4751]: I0316 00:16:03.834414 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560336-pvk5d" Mar 16 00:16:03 crc kubenswrapper[4751]: I0316 00:16:03.868912 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxh57\" (UniqueName: \"kubernetes.io/projected/d9713fca-2f31-4266-809a-76c2085543ae-kube-api-access-rxh57\") pod \"d9713fca-2f31-4266-809a-76c2085543ae\" (UID: \"d9713fca-2f31-4266-809a-76c2085543ae\") " Mar 16 00:16:03 crc kubenswrapper[4751]: I0316 00:16:03.883381 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9713fca-2f31-4266-809a-76c2085543ae-kube-api-access-rxh57" (OuterVolumeSpecName: "kube-api-access-rxh57") pod "d9713fca-2f31-4266-809a-76c2085543ae" (UID: "d9713fca-2f31-4266-809a-76c2085543ae"). InnerVolumeSpecName "kube-api-access-rxh57". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:16:03 crc kubenswrapper[4751]: I0316 00:16:03.970628 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxh57\" (UniqueName: \"kubernetes.io/projected/d9713fca-2f31-4266-809a-76c2085543ae-kube-api-access-rxh57\") on node \"crc\" DevicePath \"\"" Mar 16 00:16:04 crc kubenswrapper[4751]: I0316 00:16:04.549373 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560336-pvk5d" event={"ID":"d9713fca-2f31-4266-809a-76c2085543ae","Type":"ContainerDied","Data":"373fe23b093afb2aa9921040f5da918e5bbb024cff1761b15af76b793f48cf26"} Mar 16 00:16:04 crc kubenswrapper[4751]: I0316 00:16:04.549425 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560336-pvk5d" Mar 16 00:16:04 crc kubenswrapper[4751]: I0316 00:16:04.549429 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="373fe23b093afb2aa9921040f5da918e5bbb024cff1761b15af76b793f48cf26" Mar 16 00:16:04 crc kubenswrapper[4751]: I0316 00:16:04.919392 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29560330-lnjc4"] Mar 16 00:16:04 crc kubenswrapper[4751]: I0316 00:16:04.926474 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29560330-lnjc4"] Mar 16 00:16:06 crc kubenswrapper[4751]: I0316 00:16:06.469597 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b44b74-5471-4589-83da-e40698b419f0" path="/var/lib/kubelet/pods/f4b44b74-5471-4589-83da-e40698b419f0/volumes" Mar 16 00:17:12 crc kubenswrapper[4751]: I0316 00:17:12.920980 4751 scope.go:117] "RemoveContainer" containerID="ea3548a4dabf550b17c81a439a332770418cf213f78f87e6ac23ec95f317f195" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.514042 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fsd7v"] Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.517934 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovn-controller" containerID="cri-o://36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377" gracePeriod=30 Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.517946 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="nbdb" containerID="cri-o://a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2" gracePeriod=30 Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.518023 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="sbdb" containerID="cri-o://6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7" gracePeriod=30 Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.518066 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501" gracePeriod=30 Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.518161 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="northd" containerID="cri-o://5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a" gracePeriod=30 Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.518187 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kube-rbac-proxy-node" containerID="cri-o://73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff" gracePeriod=30 Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.518207 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovn-acl-logging" containerID="cri-o://eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51" gracePeriod=30 Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.565269 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovnkube-controller" containerID="cri-o://4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9" gracePeriod=30 Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.872633 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fsd7v_3e828458-13b3-4c40-a29d-849625254883/ovn-acl-logging/0.log" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.873314 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fsd7v_3e828458-13b3-4c40-a29d-849625254883/ovn-controller/0.log" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.873744 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933399 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9rpks"] Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933640 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="sbdb" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933653 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="sbdb" Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933684 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kube-rbac-proxy-ovn-metrics" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933691 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kube-rbac-proxy-ovn-metrics" Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933703 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovnkube-controller" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933709 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovnkube-controller" Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933722 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovn-acl-logging" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933728 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovn-acl-logging" Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933734 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="northd" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933739 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="northd" Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933765 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovn-controller" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933772 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovn-controller" Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933780 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9713fca-2f31-4266-809a-76c2085543ae" containerName="oc" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933785 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9713fca-2f31-4266-809a-76c2085543ae" containerName="oc" Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933793 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="nbdb" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933798 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="nbdb" Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933805 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kube-rbac-proxy-node" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933811 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kube-rbac-proxy-node" Mar 16 00:17:25 crc kubenswrapper[4751]: E0316 00:17:25.933836 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kubecfg-setup" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933844 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kubecfg-setup" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933951 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9713fca-2f31-4266-809a-76c2085543ae" containerName="oc" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933961 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kube-rbac-proxy-ovn-metrics" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933969 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="northd" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.933975 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovnkube-controller" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.934001 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="nbdb" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.934008 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovn-acl-logging" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.934014 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="ovn-controller" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.934022 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="sbdb" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.934030 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e828458-13b3-4c40-a29d-849625254883" containerName="kube-rbac-proxy-node" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935374 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-env-overrides\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935414 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-kubelet\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935432 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-netd\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935448 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-netns\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935476 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-script-lib\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935520 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-ovn\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935554 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e828458-13b3-4c40-a29d-849625254883-ovn-node-metrics-cert\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935567 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-openvswitch\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935580 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-var-lib-openvswitch\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935611 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-systemd-units\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935637 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935662 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-etc-openvswitch\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935687 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-systemd\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935705 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbzpx\" (UniqueName: \"kubernetes.io/projected/3e828458-13b3-4c40-a29d-849625254883-kube-api-access-fbzpx\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935725 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-node-log\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935743 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-bin\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935760 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-config\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935780 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-log-socket\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935794 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-ovn-kubernetes\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935814 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-slash\") pod \"3e828458-13b3-4c40-a29d-849625254883\" (UID: \"3e828458-13b3-4c40-a29d-849625254883\") " Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.935913 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936002 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-slash" (OuterVolumeSpecName: "host-slash") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936409 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936433 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936450 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936465 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936529 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936561 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936645 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936668 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936687 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-log-socket" (OuterVolumeSpecName: "log-socket") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936739 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936934 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936939 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936987 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.936998 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.937039 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-node-log" (OuterVolumeSpecName: "node-log") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.937548 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.942170 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e828458-13b3-4c40-a29d-849625254883-kube-api-access-fbzpx" (OuterVolumeSpecName: "kube-api-access-fbzpx") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "kube-api-access-fbzpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.946374 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e828458-13b3-4c40-a29d-849625254883-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:17:25 crc kubenswrapper[4751]: I0316 00:17:25.951872 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "3e828458-13b3-4c40-a29d-849625254883" (UID: "3e828458-13b3-4c40-a29d-849625254883"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037457 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-node-log\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037500 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8690313f-e2c6-4b1c-b725-23316db2715e-ovnkube-config\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037522 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-etc-openvswitch\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037542 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037565 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-run-ovn-kubernetes\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037584 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-run-netns\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037606 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-var-lib-openvswitch\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037622 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-run-systemd\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037646 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-cni-bin\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037744 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-log-socket\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037765 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-systemd-units\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037778 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8690313f-e2c6-4b1c-b725-23316db2715e-env-overrides\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037797 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxcwh\" (UniqueName: \"kubernetes.io/projected/8690313f-e2c6-4b1c-b725-23316db2715e-kube-api-access-vxcwh\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037815 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-cni-netd\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037836 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-run-ovn\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037887 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8690313f-e2c6-4b1c-b725-23316db2715e-ovn-node-metrics-cert\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037947 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-kubelet\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037961 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8690313f-e2c6-4b1c-b725-23316db2715e-ovnkube-script-lib\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037977 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-slash\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.037993 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-run-openvswitch\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038047 4751 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e828458-13b3-4c40-a29d-849625254883-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038058 4751 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038067 4751 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038078 4751 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038087 4751 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038096 4751 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038123 4751 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038132 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbzpx\" (UniqueName: \"kubernetes.io/projected/3e828458-13b3-4c40-a29d-849625254883-kube-api-access-fbzpx\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038140 4751 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-node-log\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038149 4751 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038157 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038164 4751 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-log-socket\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038172 4751 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038179 4751 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-slash\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038186 4751 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038195 4751 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038202 4751 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038210 4751 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038218 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e828458-13b3-4c40-a29d-849625254883-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.038226 4751 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e828458-13b3-4c40-a29d-849625254883-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.138941 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-slash\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139001 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-run-openvswitch\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139049 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-node-log\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139089 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8690313f-e2c6-4b1c-b725-23316db2715e-ovnkube-config\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139192 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-slash\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139270 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-node-log\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139223 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-etc-openvswitch\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139313 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-etc-openvswitch\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139317 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-run-openvswitch\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139396 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139369 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139544 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-run-ovn-kubernetes\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139644 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-run-netns\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139693 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-var-lib-openvswitch\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139745 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-run-systemd\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139832 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-cni-bin\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139925 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-log-socket\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.139987 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-systemd-units\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.140039 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8690313f-e2c6-4b1c-b725-23316db2715e-env-overrides\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.140193 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxcwh\" (UniqueName: \"kubernetes.io/projected/8690313f-e2c6-4b1c-b725-23316db2715e-kube-api-access-vxcwh\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.140302 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-cni-netd\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.140406 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-run-ovn\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.140470 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8690313f-e2c6-4b1c-b725-23316db2715e-ovn-node-metrics-cert\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.140659 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-kubelet\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.140712 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8690313f-e2c6-4b1c-b725-23316db2715e-ovnkube-script-lib\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.140930 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8690313f-e2c6-4b1c-b725-23316db2715e-ovnkube-config\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.141057 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-systemd-units\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.141074 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-run-ovn-kubernetes\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.141214 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-cni-netd\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.141264 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-var-lib-openvswitch\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.141305 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-cni-bin\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.141315 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-log-socket\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.141214 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-run-netns\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.141270 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-run-systemd\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.142089 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8690313f-e2c6-4b1c-b725-23316db2715e-env-overrides\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.142239 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-host-kubelet\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.142315 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8690313f-e2c6-4b1c-b725-23316db2715e-run-ovn\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.142314 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8690313f-e2c6-4b1c-b725-23316db2715e-ovnkube-script-lib\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.146415 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pdtst_caf1b5ff-47c0-474b-8429-96f4881a94c7/kube-multus/0.log" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.146505 4751 generic.go:334] "Generic (PLEG): container finished" podID="caf1b5ff-47c0-474b-8429-96f4881a94c7" containerID="8aa33b1d84d61c1f740aa738253ac05657a70c1515da2181b9fe9441b895e0eb" exitCode=2 Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.146691 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pdtst" event={"ID":"caf1b5ff-47c0-474b-8429-96f4881a94c7","Type":"ContainerDied","Data":"8aa33b1d84d61c1f740aa738253ac05657a70c1515da2181b9fe9441b895e0eb"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.147489 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8690313f-e2c6-4b1c-b725-23316db2715e-ovn-node-metrics-cert\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.147539 4751 scope.go:117] "RemoveContainer" containerID="8aa33b1d84d61c1f740aa738253ac05657a70c1515da2181b9fe9441b895e0eb" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.157772 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fsd7v_3e828458-13b3-4c40-a29d-849625254883/ovn-acl-logging/0.log" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.158954 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fsd7v_3e828458-13b3-4c40-a29d-849625254883/ovn-controller/0.log" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160241 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e828458-13b3-4c40-a29d-849625254883" containerID="4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9" exitCode=0 Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160290 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e828458-13b3-4c40-a29d-849625254883" containerID="6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7" exitCode=0 Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160315 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e828458-13b3-4c40-a29d-849625254883" containerID="a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2" exitCode=0 Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160339 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e828458-13b3-4c40-a29d-849625254883" containerID="5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a" exitCode=0 Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160360 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e828458-13b3-4c40-a29d-849625254883" containerID="880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501" exitCode=0 Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160364 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160406 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160380 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e828458-13b3-4c40-a29d-849625254883" containerID="73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff" exitCode=0 Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160480 4751 scope.go:117] "RemoveContainer" containerID="4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160497 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e828458-13b3-4c40-a29d-849625254883" containerID="eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51" exitCode=143 Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160530 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e828458-13b3-4c40-a29d-849625254883" containerID="36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377" exitCode=143 Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160454 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160600 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160634 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160665 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160697 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160725 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160749 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160765 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160788 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160811 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160828 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160844 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160860 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160874 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160889 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160904 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160918 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160932 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160953 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.160982 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161001 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161018 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161032 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161047 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161060 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161075 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161088 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161168 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161197 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fsd7v" event={"ID":"3e828458-13b3-4c40-a29d-849625254883","Type":"ContainerDied","Data":"cec470ab0cd1126c195fed32aad8163cdfba13ca4374503b921df69ffafe7a28"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161222 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161240 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161254 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161268 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161284 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161298 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161311 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161326 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.161342 4751 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca"} Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.175582 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxcwh\" (UniqueName: \"kubernetes.io/projected/8690313f-e2c6-4b1c-b725-23316db2715e-kube-api-access-vxcwh\") pod \"ovnkube-node-9rpks\" (UID: \"8690313f-e2c6-4b1c-b725-23316db2715e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.198853 4751 scope.go:117] "RemoveContainer" containerID="6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.222699 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fsd7v"] Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.230799 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fsd7v"] Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.237141 4751 scope.go:117] "RemoveContainer" containerID="a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.258354 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.261797 4751 scope.go:117] "RemoveContainer" containerID="5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.279942 4751 scope.go:117] "RemoveContainer" containerID="880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501" Mar 16 00:17:26 crc kubenswrapper[4751]: W0316 00:17:26.288544 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8690313f_e2c6_4b1c_b725_23316db2715e.slice/crio-237f62c763e3b46ed2b05f518043435f43ef210ea6657442d50857d8f72c20bd WatchSource:0}: Error finding container 237f62c763e3b46ed2b05f518043435f43ef210ea6657442d50857d8f72c20bd: Status 404 returned error can't find the container with id 237f62c763e3b46ed2b05f518043435f43ef210ea6657442d50857d8f72c20bd Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.297197 4751 scope.go:117] "RemoveContainer" containerID="73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.324652 4751 scope.go:117] "RemoveContainer" containerID="eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.340302 4751 scope.go:117] "RemoveContainer" containerID="36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.361991 4751 scope.go:117] "RemoveContainer" containerID="1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.390423 4751 scope.go:117] "RemoveContainer" containerID="4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9" Mar 16 00:17:26 crc kubenswrapper[4751]: E0316 00:17:26.390864 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": container with ID starting with 4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9 not found: ID does not exist" containerID="4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.390942 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} err="failed to get container status \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": rpc error: code = NotFound desc = could not find container \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": container with ID starting with 4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.390994 4751 scope.go:117] "RemoveContainer" containerID="6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7" Mar 16 00:17:26 crc kubenswrapper[4751]: E0316 00:17:26.391528 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": container with ID starting with 6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7 not found: ID does not exist" containerID="6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.391574 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} err="failed to get container status \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": rpc error: code = NotFound desc = could not find container \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": container with ID starting with 6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.391608 4751 scope.go:117] "RemoveContainer" containerID="a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2" Mar 16 00:17:26 crc kubenswrapper[4751]: E0316 00:17:26.391949 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": container with ID starting with a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2 not found: ID does not exist" containerID="a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.392000 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} err="failed to get container status \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": rpc error: code = NotFound desc = could not find container \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": container with ID starting with a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.392035 4751 scope.go:117] "RemoveContainer" containerID="5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a" Mar 16 00:17:26 crc kubenswrapper[4751]: E0316 00:17:26.392408 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": container with ID starting with 5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a not found: ID does not exist" containerID="5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.392441 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} err="failed to get container status \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": rpc error: code = NotFound desc = could not find container \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": container with ID starting with 5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.392469 4751 scope.go:117] "RemoveContainer" containerID="880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501" Mar 16 00:17:26 crc kubenswrapper[4751]: E0316 00:17:26.392774 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": container with ID starting with 880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501 not found: ID does not exist" containerID="880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.392806 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} err="failed to get container status \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": rpc error: code = NotFound desc = could not find container \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": container with ID starting with 880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.392824 4751 scope.go:117] "RemoveContainer" containerID="73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff" Mar 16 00:17:26 crc kubenswrapper[4751]: E0316 00:17:26.393125 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": container with ID starting with 73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff not found: ID does not exist" containerID="73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.393155 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} err="failed to get container status \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": rpc error: code = NotFound desc = could not find container \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": container with ID starting with 73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.393175 4751 scope.go:117] "RemoveContainer" containerID="eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51" Mar 16 00:17:26 crc kubenswrapper[4751]: E0316 00:17:26.393577 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51\": container with ID starting with eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51 not found: ID does not exist" containerID="eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.393610 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} err="failed to get container status \"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51\": rpc error: code = NotFound desc = could not find container \"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51\": container with ID starting with eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.393638 4751 scope.go:117] "RemoveContainer" containerID="36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377" Mar 16 00:17:26 crc kubenswrapper[4751]: E0316 00:17:26.394020 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377\": container with ID starting with 36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377 not found: ID does not exist" containerID="36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.394049 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} err="failed to get container status \"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377\": rpc error: code = NotFound desc = could not find container \"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377\": container with ID starting with 36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.394073 4751 scope.go:117] "RemoveContainer" containerID="1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca" Mar 16 00:17:26 crc kubenswrapper[4751]: E0316 00:17:26.394573 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca\": container with ID starting with 1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca not found: ID does not exist" containerID="1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.394627 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca"} err="failed to get container status \"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca\": rpc error: code = NotFound desc = could not find container \"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca\": container with ID starting with 1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.394644 4751 scope.go:117] "RemoveContainer" containerID="4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.395003 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} err="failed to get container status \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": rpc error: code = NotFound desc = could not find container \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": container with ID starting with 4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.395030 4751 scope.go:117] "RemoveContainer" containerID="6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.395338 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} err="failed to get container status \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": rpc error: code = NotFound desc = could not find container \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": container with ID starting with 6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.395369 4751 scope.go:117] "RemoveContainer" containerID="a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.395782 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} err="failed to get container status \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": rpc error: code = NotFound desc = could not find container \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": container with ID starting with a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.395806 4751 scope.go:117] "RemoveContainer" containerID="5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.396198 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} err="failed to get container status \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": rpc error: code = NotFound desc = could not find container \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": container with ID starting with 5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.396230 4751 scope.go:117] "RemoveContainer" containerID="880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.396664 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} err="failed to get container status \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": rpc error: code = NotFound desc = could not find container \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": container with ID starting with 880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.396710 4751 scope.go:117] "RemoveContainer" containerID="73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.397152 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} err="failed to get container status \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": rpc error: code = NotFound desc = could not find container \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": container with ID starting with 73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.397200 4751 scope.go:117] "RemoveContainer" containerID="eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.397494 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} err="failed to get container status \"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51\": rpc error: code = NotFound desc = could not find container \"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51\": container with ID starting with eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.397521 4751 scope.go:117] "RemoveContainer" containerID="36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.397977 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} err="failed to get container status \"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377\": rpc error: code = NotFound desc = could not find container \"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377\": container with ID starting with 36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.398010 4751 scope.go:117] "RemoveContainer" containerID="1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.398578 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca"} err="failed to get container status \"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca\": rpc error: code = NotFound desc = could not find container \"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca\": container with ID starting with 1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.398611 4751 scope.go:117] "RemoveContainer" containerID="4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.398958 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} err="failed to get container status \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": rpc error: code = NotFound desc = could not find container \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": container with ID starting with 4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.399006 4751 scope.go:117] "RemoveContainer" containerID="6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.399306 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} err="failed to get container status \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": rpc error: code = NotFound desc = could not find container \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": container with ID starting with 6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.399334 4751 scope.go:117] "RemoveContainer" containerID="a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.399690 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} err="failed to get container status \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": rpc error: code = NotFound desc = could not find container \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": container with ID starting with a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.399737 4751 scope.go:117] "RemoveContainer" containerID="5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.400092 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} err="failed to get container status \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": rpc error: code = NotFound desc = could not find container \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": container with ID starting with 5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.400161 4751 scope.go:117] "RemoveContainer" containerID="880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.400458 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} err="failed to get container status \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": rpc error: code = NotFound desc = could not find container \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": container with ID starting with 880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.400483 4751 scope.go:117] "RemoveContainer" containerID="73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.400806 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} err="failed to get container status \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": rpc error: code = NotFound desc = could not find container \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": container with ID starting with 73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.400833 4751 scope.go:117] "RemoveContainer" containerID="eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.401159 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} err="failed to get container status \"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51\": rpc error: code = NotFound desc = could not find container \"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51\": container with ID starting with eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.401191 4751 scope.go:117] "RemoveContainer" containerID="36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.401576 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} err="failed to get container status \"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377\": rpc error: code = NotFound desc = could not find container \"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377\": container with ID starting with 36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.401623 4751 scope.go:117] "RemoveContainer" containerID="1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.402036 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca"} err="failed to get container status \"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca\": rpc error: code = NotFound desc = could not find container \"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca\": container with ID starting with 1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.402063 4751 scope.go:117] "RemoveContainer" containerID="4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.402376 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} err="failed to get container status \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": rpc error: code = NotFound desc = could not find container \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": container with ID starting with 4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.402425 4751 scope.go:117] "RemoveContainer" containerID="6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.402711 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} err="failed to get container status \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": rpc error: code = NotFound desc = could not find container \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": container with ID starting with 6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.402739 4751 scope.go:117] "RemoveContainer" containerID="a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.403131 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} err="failed to get container status \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": rpc error: code = NotFound desc = could not find container \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": container with ID starting with a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.403170 4751 scope.go:117] "RemoveContainer" containerID="5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.403536 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} err="failed to get container status \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": rpc error: code = NotFound desc = could not find container \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": container with ID starting with 5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.403557 4751 scope.go:117] "RemoveContainer" containerID="880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.403837 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} err="failed to get container status \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": rpc error: code = NotFound desc = could not find container \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": container with ID starting with 880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.403868 4751 scope.go:117] "RemoveContainer" containerID="73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.404194 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} err="failed to get container status \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": rpc error: code = NotFound desc = could not find container \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": container with ID starting with 73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.404254 4751 scope.go:117] "RemoveContainer" containerID="eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.404560 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51"} err="failed to get container status \"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51\": rpc error: code = NotFound desc = could not find container \"eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51\": container with ID starting with eb72cc4a8b66c4dce863bb652168bb6629b9fd492ea202b67df496927c216c51 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.404588 4751 scope.go:117] "RemoveContainer" containerID="36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.404851 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377"} err="failed to get container status \"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377\": rpc error: code = NotFound desc = could not find container \"36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377\": container with ID starting with 36945ff921332645602e6016642f3951ab7f86287e44cbb67a72b1605dfce377 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.404873 4751 scope.go:117] "RemoveContainer" containerID="1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.405180 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca"} err="failed to get container status \"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca\": rpc error: code = NotFound desc = could not find container \"1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca\": container with ID starting with 1dcfdf1d767fd6024e1a61d1c566c106362bfccc62a1d5ea60714728c17d5fca not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.405207 4751 scope.go:117] "RemoveContainer" containerID="4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.405483 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9"} err="failed to get container status \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": rpc error: code = NotFound desc = could not find container \"4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9\": container with ID starting with 4e221dffa41f7367dbf452ea6a93d450a5403162173437eeca75394d705dc6b9 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.405505 4751 scope.go:117] "RemoveContainer" containerID="6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.405730 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7"} err="failed to get container status \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": rpc error: code = NotFound desc = could not find container \"6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7\": container with ID starting with 6fd43cccfbf75a9afbbc34d7e1f3603022d4befba3b9d8a77ac82e56baffc8e7 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.405759 4751 scope.go:117] "RemoveContainer" containerID="a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.406592 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2"} err="failed to get container status \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": rpc error: code = NotFound desc = could not find container \"a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2\": container with ID starting with a202035e5a04874b7e9b264f593813e9953a1f48961ececdc9166a8636dca3b2 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.406621 4751 scope.go:117] "RemoveContainer" containerID="5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.406924 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a"} err="failed to get container status \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": rpc error: code = NotFound desc = could not find container \"5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a\": container with ID starting with 5241f29cccbb249cd6d86d0592cf4e5dda9b9d0b22c98f02c9b5c286fcfaec3a not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.406945 4751 scope.go:117] "RemoveContainer" containerID="880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.407477 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501"} err="failed to get container status \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": rpc error: code = NotFound desc = could not find container \"880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501\": container with ID starting with 880bab96078e120d8ce0c504f31bca861c6552aadde54ffe3460d53e746b5501 not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.407532 4751 scope.go:117] "RemoveContainer" containerID="73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.407802 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff"} err="failed to get container status \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": rpc error: code = NotFound desc = could not find container \"73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff\": container with ID starting with 73b27e598e9296a40e5813c3c698981ab7a776eec81a3b3f6b534cc8f37eacff not found: ID does not exist" Mar 16 00:17:26 crc kubenswrapper[4751]: I0316 00:17:26.466233 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e828458-13b3-4c40-a29d-849625254883" path="/var/lib/kubelet/pods/3e828458-13b3-4c40-a29d-849625254883/volumes" Mar 16 00:17:27 crc kubenswrapper[4751]: I0316 00:17:27.173774 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pdtst_caf1b5ff-47c0-474b-8429-96f4881a94c7/kube-multus/0.log" Mar 16 00:17:27 crc kubenswrapper[4751]: I0316 00:17:27.175469 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pdtst" event={"ID":"caf1b5ff-47c0-474b-8429-96f4881a94c7","Type":"ContainerStarted","Data":"5e900628dd19b32cbb89f5fca082823c47f41a4e7ef4abf766e43a33c48f5be2"} Mar 16 00:17:27 crc kubenswrapper[4751]: I0316 00:17:27.177901 4751 generic.go:334] "Generic (PLEG): container finished" podID="8690313f-e2c6-4b1c-b725-23316db2715e" containerID="ad3b8e518170b2c27a437446b809060cce5a9af9af12c496fcd84852871868ed" exitCode=0 Mar 16 00:17:27 crc kubenswrapper[4751]: I0316 00:17:27.177944 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerDied","Data":"ad3b8e518170b2c27a437446b809060cce5a9af9af12c496fcd84852871868ed"} Mar 16 00:17:27 crc kubenswrapper[4751]: I0316 00:17:27.177988 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerStarted","Data":"237f62c763e3b46ed2b05f518043435f43ef210ea6657442d50857d8f72c20bd"} Mar 16 00:17:27 crc kubenswrapper[4751]: I0316 00:17:27.854732 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:17:27 crc kubenswrapper[4751]: I0316 00:17:27.855138 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:17:28 crc kubenswrapper[4751]: I0316 00:17:28.188624 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerStarted","Data":"59f5aad6bf549a2bbbd1d0ce8d09b605175c9973731fdd3043012aee6487ff75"} Mar 16 00:17:28 crc kubenswrapper[4751]: I0316 00:17:28.188670 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerStarted","Data":"5f19d511a4e3c99d7ffd8b694cb2792787ba3e287fa1cd6544c6ed4f30183b16"} Mar 16 00:17:28 crc kubenswrapper[4751]: I0316 00:17:28.188712 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerStarted","Data":"431b99c9b2cb98b1cb9b190f4371c32091e17e901e39d179b4008b012e4c241d"} Mar 16 00:17:28 crc kubenswrapper[4751]: I0316 00:17:28.188726 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerStarted","Data":"03e87eed52d013f72dfa7b35ca287a15086ce52757cc76437a38c33d5c491ce8"} Mar 16 00:17:28 crc kubenswrapper[4751]: I0316 00:17:28.188738 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerStarted","Data":"33b403851e44e7085ce2679bd4e8cf4d8786e84de1e069c526d513a1195fcc07"} Mar 16 00:17:28 crc kubenswrapper[4751]: I0316 00:17:28.188751 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerStarted","Data":"b1db3d5329daa5531def5f8526142f30ae0bb359539a1e76fa807c1f5fac9ada"} Mar 16 00:17:31 crc kubenswrapper[4751]: I0316 00:17:31.236899 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerStarted","Data":"a0ddb1cfe930ae22199767010ad11bd357461adec4d86d68cfeb8108d9da36b2"} Mar 16 00:17:33 crc kubenswrapper[4751]: I0316 00:17:33.252267 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" event={"ID":"8690313f-e2c6-4b1c-b725-23316db2715e","Type":"ContainerStarted","Data":"d69b169335b819c9ca5ebb3df65996efe6b780a0c1d096c83c50fca6c24484f4"} Mar 16 00:17:33 crc kubenswrapper[4751]: I0316 00:17:33.254218 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:33 crc kubenswrapper[4751]: I0316 00:17:33.287608 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:33 crc kubenswrapper[4751]: I0316 00:17:33.291146 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" podStartSLOduration=8.291133442 podStartE2EDuration="8.291133442s" podCreationTimestamp="2026-03-16 00:17:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:17:33.288155598 +0000 UTC m=+681.557487888" watchObservedRunningTime="2026-03-16 00:17:33.291133442 +0000 UTC m=+681.560465732" Mar 16 00:17:34 crc kubenswrapper[4751]: I0316 00:17:34.261848 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:34 crc kubenswrapper[4751]: I0316 00:17:34.261909 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:34 crc kubenswrapper[4751]: I0316 00:17:34.304615 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:56 crc kubenswrapper[4751]: I0316 00:17:56.295524 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9rpks" Mar 16 00:17:59 crc kubenswrapper[4751]: I0316 00:17:59.612425 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:17:59 crc kubenswrapper[4751]: I0316 00:17:59.612774 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.151512 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560338-ghpdr"] Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.152576 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560338-ghpdr" Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.156440 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.157308 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.159190 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.161923 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560338-ghpdr"] Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.221072 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdznz\" (UniqueName: \"kubernetes.io/projected/a2b5d3b3-849a-4d38-b757-fe5f36befaa5-kube-api-access-kdznz\") pod \"auto-csr-approver-29560338-ghpdr\" (UID: \"a2b5d3b3-849a-4d38-b757-fe5f36befaa5\") " pod="openshift-infra/auto-csr-approver-29560338-ghpdr" Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.322444 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdznz\" (UniqueName: \"kubernetes.io/projected/a2b5d3b3-849a-4d38-b757-fe5f36befaa5-kube-api-access-kdznz\") pod \"auto-csr-approver-29560338-ghpdr\" (UID: \"a2b5d3b3-849a-4d38-b757-fe5f36befaa5\") " pod="openshift-infra/auto-csr-approver-29560338-ghpdr" Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.356150 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdznz\" (UniqueName: \"kubernetes.io/projected/a2b5d3b3-849a-4d38-b757-fe5f36befaa5-kube-api-access-kdznz\") pod \"auto-csr-approver-29560338-ghpdr\" (UID: \"a2b5d3b3-849a-4d38-b757-fe5f36befaa5\") " pod="openshift-infra/auto-csr-approver-29560338-ghpdr" Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.474910 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560338-ghpdr" Mar 16 00:18:00 crc kubenswrapper[4751]: I0316 00:18:00.746069 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560338-ghpdr"] Mar 16 00:18:01 crc kubenswrapper[4751]: I0316 00:18:01.658590 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560338-ghpdr" event={"ID":"a2b5d3b3-849a-4d38-b757-fe5f36befaa5","Type":"ContainerStarted","Data":"cd3161d82ed9200651a23a5d6dc5a20cc019fee8261fe3587c5c8b4243128aca"} Mar 16 00:18:02 crc kubenswrapper[4751]: I0316 00:18:02.667892 4751 generic.go:334] "Generic (PLEG): container finished" podID="a2b5d3b3-849a-4d38-b757-fe5f36befaa5" containerID="89df761e05427df967335e82e301b66d2caf73c0911761cb993a5e767380b188" exitCode=0 Mar 16 00:18:02 crc kubenswrapper[4751]: I0316 00:18:02.667946 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560338-ghpdr" event={"ID":"a2b5d3b3-849a-4d38-b757-fe5f36befaa5","Type":"ContainerDied","Data":"89df761e05427df967335e82e301b66d2caf73c0911761cb993a5e767380b188"} Mar 16 00:18:04 crc kubenswrapper[4751]: I0316 00:18:04.046917 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560338-ghpdr" Mar 16 00:18:04 crc kubenswrapper[4751]: I0316 00:18:04.074455 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdznz\" (UniqueName: \"kubernetes.io/projected/a2b5d3b3-849a-4d38-b757-fe5f36befaa5-kube-api-access-kdznz\") pod \"a2b5d3b3-849a-4d38-b757-fe5f36befaa5\" (UID: \"a2b5d3b3-849a-4d38-b757-fe5f36befaa5\") " Mar 16 00:18:04 crc kubenswrapper[4751]: I0316 00:18:04.080571 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b5d3b3-849a-4d38-b757-fe5f36befaa5-kube-api-access-kdznz" (OuterVolumeSpecName: "kube-api-access-kdznz") pod "a2b5d3b3-849a-4d38-b757-fe5f36befaa5" (UID: "a2b5d3b3-849a-4d38-b757-fe5f36befaa5"). InnerVolumeSpecName "kube-api-access-kdznz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:18:04 crc kubenswrapper[4751]: I0316 00:18:04.176477 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdznz\" (UniqueName: \"kubernetes.io/projected/a2b5d3b3-849a-4d38-b757-fe5f36befaa5-kube-api-access-kdznz\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:04 crc kubenswrapper[4751]: I0316 00:18:04.683362 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560338-ghpdr" event={"ID":"a2b5d3b3-849a-4d38-b757-fe5f36befaa5","Type":"ContainerDied","Data":"cd3161d82ed9200651a23a5d6dc5a20cc019fee8261fe3587c5c8b4243128aca"} Mar 16 00:18:04 crc kubenswrapper[4751]: I0316 00:18:04.683426 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd3161d82ed9200651a23a5d6dc5a20cc019fee8261fe3587c5c8b4243128aca" Mar 16 00:18:04 crc kubenswrapper[4751]: I0316 00:18:04.683444 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560338-ghpdr" Mar 16 00:18:05 crc kubenswrapper[4751]: I0316 00:18:05.123597 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29560332-vtxlh"] Mar 16 00:18:05 crc kubenswrapper[4751]: I0316 00:18:05.145579 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29560332-vtxlh"] Mar 16 00:18:06 crc kubenswrapper[4751]: I0316 00:18:06.466865 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd" path="/var/lib/kubelet/pods/44acbb6e-1b8e-491e-bb5f-2b6a5090dfcd/volumes" Mar 16 00:18:13 crc kubenswrapper[4751]: I0316 00:18:13.015593 4751 scope.go:117] "RemoveContainer" containerID="243ee17a8b9df926b7fee9bc4ee3b69249c33eb51fe3c27452a104ef90b678b2" Mar 16 00:18:24 crc kubenswrapper[4751]: I0316 00:18:24.791614 4751 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 16 00:18:27 crc kubenswrapper[4751]: I0316 00:18:27.855199 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:18:27 crc kubenswrapper[4751]: I0316 00:18:27.855632 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:18:27 crc kubenswrapper[4751]: I0316 00:18:27.855684 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:18:27 crc kubenswrapper[4751]: I0316 00:18:27.856361 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a45820bd88f0ba2e569a216d14a8ef916b60532a9717b18faa5c4a39cc910750"} pod="openshift-machine-config-operator/machine-config-daemon-6fccz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 16 00:18:27 crc kubenswrapper[4751]: I0316 00:18:27.856425 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" containerID="cri-o://a45820bd88f0ba2e569a216d14a8ef916b60532a9717b18faa5c4a39cc910750" gracePeriod=600 Mar 16 00:18:28 crc kubenswrapper[4751]: I0316 00:18:28.857788 4751 generic.go:334] "Generic (PLEG): container finished" podID="25429719-d18d-4909-8455-701d09423c36" containerID="a45820bd88f0ba2e569a216d14a8ef916b60532a9717b18faa5c4a39cc910750" exitCode=0 Mar 16 00:18:28 crc kubenswrapper[4751]: I0316 00:18:28.857835 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerDied","Data":"a45820bd88f0ba2e569a216d14a8ef916b60532a9717b18faa5c4a39cc910750"} Mar 16 00:18:28 crc kubenswrapper[4751]: I0316 00:18:28.858085 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerStarted","Data":"65efad03d473120c12153253332bfeb08a8e6a82c87fa93d5e003a10153418a4"} Mar 16 00:18:28 crc kubenswrapper[4751]: I0316 00:18:28.858131 4751 scope.go:117] "RemoveContainer" containerID="36270f1e57b171978cc55e8e0d2e8251ea29ff34e7dce546d8c65c5b8b799266" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.286703 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mp6q2"] Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.287208 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mp6q2" podUID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerName="registry-server" containerID="cri-o://85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66" gracePeriod=30 Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.679796 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.754254 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5wrv\" (UniqueName: \"kubernetes.io/projected/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-kube-api-access-r5wrv\") pod \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.754341 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-utilities\") pod \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.754368 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-catalog-content\") pod \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\" (UID: \"3bd04a02-93f5-4cd5-88c9-bccc3cae359f\") " Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.761219 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-utilities" (OuterVolumeSpecName: "utilities") pod "3bd04a02-93f5-4cd5-88c9-bccc3cae359f" (UID: "3bd04a02-93f5-4cd5-88c9-bccc3cae359f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.766656 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-kube-api-access-r5wrv" (OuterVolumeSpecName: "kube-api-access-r5wrv") pod "3bd04a02-93f5-4cd5-88c9-bccc3cae359f" (UID: "3bd04a02-93f5-4cd5-88c9-bccc3cae359f"). InnerVolumeSpecName "kube-api-access-r5wrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.792800 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bd04a02-93f5-4cd5-88c9-bccc3cae359f" (UID: "3bd04a02-93f5-4cd5-88c9-bccc3cae359f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.855387 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5wrv\" (UniqueName: \"kubernetes.io/projected/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-kube-api-access-r5wrv\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.855436 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.855450 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd04a02-93f5-4cd5-88c9-bccc3cae359f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.868823 4751 generic.go:334] "Generic (PLEG): container finished" podID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerID="85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66" exitCode=0 Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.868932 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mp6q2" event={"ID":"3bd04a02-93f5-4cd5-88c9-bccc3cae359f","Type":"ContainerDied","Data":"85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66"} Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.868996 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mp6q2" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.869017 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mp6q2" event={"ID":"3bd04a02-93f5-4cd5-88c9-bccc3cae359f","Type":"ContainerDied","Data":"b0f9fae330d209d0b9963d885a2feae7f5e9ea500fea40a06032e74af6fc2853"} Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.869076 4751 scope.go:117] "RemoveContainer" containerID="85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.894755 4751 scope.go:117] "RemoveContainer" containerID="32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.909841 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mp6q2"] Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.918186 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mp6q2"] Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.933562 4751 scope.go:117] "RemoveContainer" containerID="cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.958673 4751 scope.go:117] "RemoveContainer" containerID="85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66" Mar 16 00:18:29 crc kubenswrapper[4751]: E0316 00:18:29.959513 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66\": container with ID starting with 85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66 not found: ID does not exist" containerID="85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.959580 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66"} err="failed to get container status \"85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66\": rpc error: code = NotFound desc = could not find container \"85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66\": container with ID starting with 85fc973ea64429b8efc877ce9fcc1e59ced245f75f0effe743f4de11d16a2c66 not found: ID does not exist" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.959627 4751 scope.go:117] "RemoveContainer" containerID="32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5" Mar 16 00:18:29 crc kubenswrapper[4751]: E0316 00:18:29.960299 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5\": container with ID starting with 32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5 not found: ID does not exist" containerID="32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.960364 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5"} err="failed to get container status \"32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5\": rpc error: code = NotFound desc = could not find container \"32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5\": container with ID starting with 32209f26481bf6906bf6d200ce2eb9a0186786e9c54a0b1d533ecd37820e28a5 not found: ID does not exist" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.960406 4751 scope.go:117] "RemoveContainer" containerID="cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca" Mar 16 00:18:29 crc kubenswrapper[4751]: E0316 00:18:29.960910 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca\": container with ID starting with cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca not found: ID does not exist" containerID="cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca" Mar 16 00:18:29 crc kubenswrapper[4751]: I0316 00:18:29.960953 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca"} err="failed to get container status \"cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca\": rpc error: code = NotFound desc = could not find container \"cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca\": container with ID starting with cea4f9bbadd4e1e6e3591a3a442dcaba6ebeb6826e09c97f4e2a273de7c77eca not found: ID does not exist" Mar 16 00:18:30 crc kubenswrapper[4751]: I0316 00:18:30.462296 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" path="/var/lib/kubelet/pods/3bd04a02-93f5-4cd5-88c9-bccc3cae359f/volumes" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.030424 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4"] Mar 16 00:18:33 crc kubenswrapper[4751]: E0316 00:18:33.031676 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerName="extract-content" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.031709 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerName="extract-content" Mar 16 00:18:33 crc kubenswrapper[4751]: E0316 00:18:33.031747 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b5d3b3-849a-4d38-b757-fe5f36befaa5" containerName="oc" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.031766 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b5d3b3-849a-4d38-b757-fe5f36befaa5" containerName="oc" Mar 16 00:18:33 crc kubenswrapper[4751]: E0316 00:18:33.031801 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerName="extract-utilities" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.031820 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerName="extract-utilities" Mar 16 00:18:33 crc kubenswrapper[4751]: E0316 00:18:33.031875 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerName="registry-server" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.031895 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerName="registry-server" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.032176 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bd04a02-93f5-4cd5-88c9-bccc3cae359f" containerName="registry-server" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.032221 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b5d3b3-849a-4d38-b757-fe5f36befaa5" containerName="oc" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.033933 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.036391 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.062190 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4"] Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.193364 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.193448 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.193577 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9mzw\" (UniqueName: \"kubernetes.io/projected/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-kube-api-access-k9mzw\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.294163 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.294211 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.294247 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9mzw\" (UniqueName: \"kubernetes.io/projected/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-kube-api-access-k9mzw\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.295198 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.295520 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.328917 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9mzw\" (UniqueName: \"kubernetes.io/projected/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-kube-api-access-k9mzw\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.357082 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.670849 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4"] Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.901128 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" event={"ID":"bd8ea216-286a-4575-8b3e-fd8c3f2afcda","Type":"ContainerStarted","Data":"9384eb4d59210722dddd052275b777ee1f698dfb42193321e322ad72305105e9"} Mar 16 00:18:33 crc kubenswrapper[4751]: I0316 00:18:33.901174 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" event={"ID":"bd8ea216-286a-4575-8b3e-fd8c3f2afcda","Type":"ContainerStarted","Data":"1df06d886a16aa120671056396b58c178ad7f09245735c77a0469e65125587e1"} Mar 16 00:18:34 crc kubenswrapper[4751]: I0316 00:18:34.910605 4751 generic.go:334] "Generic (PLEG): container finished" podID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerID="9384eb4d59210722dddd052275b777ee1f698dfb42193321e322ad72305105e9" exitCode=0 Mar 16 00:18:34 crc kubenswrapper[4751]: I0316 00:18:34.910696 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" event={"ID":"bd8ea216-286a-4575-8b3e-fd8c3f2afcda","Type":"ContainerDied","Data":"9384eb4d59210722dddd052275b777ee1f698dfb42193321e322ad72305105e9"} Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.232522 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9qhvg"] Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.233731 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.237477 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qhvg"] Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.380126 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-utilities\") pod \"redhat-operators-9qhvg\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.380164 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh5rv\" (UniqueName: \"kubernetes.io/projected/d8f59b74-790e-44dd-828f-5fee749d88db-kube-api-access-vh5rv\") pod \"redhat-operators-9qhvg\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.380204 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-catalog-content\") pod \"redhat-operators-9qhvg\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.480776 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-utilities\") pod \"redhat-operators-9qhvg\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.481029 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh5rv\" (UniqueName: \"kubernetes.io/projected/d8f59b74-790e-44dd-828f-5fee749d88db-kube-api-access-vh5rv\") pod \"redhat-operators-9qhvg\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.481062 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-catalog-content\") pod \"redhat-operators-9qhvg\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.481344 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-utilities\") pod \"redhat-operators-9qhvg\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.481362 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-catalog-content\") pod \"redhat-operators-9qhvg\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.517211 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh5rv\" (UniqueName: \"kubernetes.io/projected/d8f59b74-790e-44dd-828f-5fee749d88db-kube-api-access-vh5rv\") pod \"redhat-operators-9qhvg\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.550769 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.751304 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qhvg"] Mar 16 00:18:36 crc kubenswrapper[4751]: W0316 00:18:36.760190 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8f59b74_790e_44dd_828f_5fee749d88db.slice/crio-58ac8595539517f1b415503d2714aaadb94ec31058dee58aba1c7744360c6f93 WatchSource:0}: Error finding container 58ac8595539517f1b415503d2714aaadb94ec31058dee58aba1c7744360c6f93: Status 404 returned error can't find the container with id 58ac8595539517f1b415503d2714aaadb94ec31058dee58aba1c7744360c6f93 Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.923343 4751 generic.go:334] "Generic (PLEG): container finished" podID="d8f59b74-790e-44dd-828f-5fee749d88db" containerID="87cb230e571d82f28acff8dfcda133d69e99b577a91f5de28596f5e60e542f94" exitCode=0 Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.923413 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qhvg" event={"ID":"d8f59b74-790e-44dd-828f-5fee749d88db","Type":"ContainerDied","Data":"87cb230e571d82f28acff8dfcda133d69e99b577a91f5de28596f5e60e542f94"} Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.923670 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qhvg" event={"ID":"d8f59b74-790e-44dd-828f-5fee749d88db","Type":"ContainerStarted","Data":"58ac8595539517f1b415503d2714aaadb94ec31058dee58aba1c7744360c6f93"} Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.926171 4751 generic.go:334] "Generic (PLEG): container finished" podID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerID="869b2f4fd96ccdc3ab1e8a48415e63830eff173fdc1eeeac42c178c0c5e19e1f" exitCode=0 Mar 16 00:18:36 crc kubenswrapper[4751]: I0316 00:18:36.926217 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" event={"ID":"bd8ea216-286a-4575-8b3e-fd8c3f2afcda","Type":"ContainerDied","Data":"869b2f4fd96ccdc3ab1e8a48415e63830eff173fdc1eeeac42c178c0c5e19e1f"} Mar 16 00:18:37 crc kubenswrapper[4751]: I0316 00:18:37.934825 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qhvg" event={"ID":"d8f59b74-790e-44dd-828f-5fee749d88db","Type":"ContainerStarted","Data":"31dd9a6f4f9d2c699906be7b1a74aebb8679e2bb4a3ea1437049eb86a56c7632"} Mar 16 00:18:37 crc kubenswrapper[4751]: I0316 00:18:37.937418 4751 generic.go:334] "Generic (PLEG): container finished" podID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerID="a0d4049893a68bdbaf7959ea07a6826e04d00cd89a79a2dbef6563752ecd72ee" exitCode=0 Mar 16 00:18:37 crc kubenswrapper[4751]: I0316 00:18:37.937880 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" event={"ID":"bd8ea216-286a-4575-8b3e-fd8c3f2afcda","Type":"ContainerDied","Data":"a0d4049893a68bdbaf7959ea07a6826e04d00cd89a79a2dbef6563752ecd72ee"} Mar 16 00:18:38 crc kubenswrapper[4751]: I0316 00:18:38.949284 4751 generic.go:334] "Generic (PLEG): container finished" podID="d8f59b74-790e-44dd-828f-5fee749d88db" containerID="31dd9a6f4f9d2c699906be7b1a74aebb8679e2bb4a3ea1437049eb86a56c7632" exitCode=0 Mar 16 00:18:38 crc kubenswrapper[4751]: I0316 00:18:38.950461 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qhvg" event={"ID":"d8f59b74-790e-44dd-828f-5fee749d88db","Type":"ContainerDied","Data":"31dd9a6f4f9d2c699906be7b1a74aebb8679e2bb4a3ea1437049eb86a56c7632"} Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.285277 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.418237 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-util\") pod \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.418339 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-bundle\") pod \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.418365 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9mzw\" (UniqueName: \"kubernetes.io/projected/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-kube-api-access-k9mzw\") pod \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\" (UID: \"bd8ea216-286a-4575-8b3e-fd8c3f2afcda\") " Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.421157 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-bundle" (OuterVolumeSpecName: "bundle") pod "bd8ea216-286a-4575-8b3e-fd8c3f2afcda" (UID: "bd8ea216-286a-4575-8b3e-fd8c3f2afcda"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.424927 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-kube-api-access-k9mzw" (OuterVolumeSpecName: "kube-api-access-k9mzw") pod "bd8ea216-286a-4575-8b3e-fd8c3f2afcda" (UID: "bd8ea216-286a-4575-8b3e-fd8c3f2afcda"). InnerVolumeSpecName "kube-api-access-k9mzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.439386 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-util" (OuterVolumeSpecName: "util") pod "bd8ea216-286a-4575-8b3e-fd8c3f2afcda" (UID: "bd8ea216-286a-4575-8b3e-fd8c3f2afcda"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.519602 4751 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-util\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.519632 4751 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.519644 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9mzw\" (UniqueName: \"kubernetes.io/projected/bd8ea216-286a-4575-8b3e-fd8c3f2afcda-kube-api-access-k9mzw\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.609890 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm"] Mar 16 00:18:39 crc kubenswrapper[4751]: E0316 00:18:39.610142 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerName="util" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.610157 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerName="util" Mar 16 00:18:39 crc kubenswrapper[4751]: E0316 00:18:39.610179 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerName="pull" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.610186 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerName="pull" Mar 16 00:18:39 crc kubenswrapper[4751]: E0316 00:18:39.610198 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerName="extract" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.610205 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerName="extract" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.610308 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd8ea216-286a-4575-8b3e-fd8c3f2afcda" containerName="extract" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.611138 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.622772 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm"] Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.721651 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.721705 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xcmx\" (UniqueName: \"kubernetes.io/projected/89cd9003-4d55-44e5-8f14-9caceb86e67a-kube-api-access-5xcmx\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.721746 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.823402 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.823485 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xcmx\" (UniqueName: \"kubernetes.io/projected/89cd9003-4d55-44e5-8f14-9caceb86e67a-kube-api-access-5xcmx\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.823548 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.823971 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.824122 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.850700 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xcmx\" (UniqueName: \"kubernetes.io/projected/89cd9003-4d55-44e5-8f14-9caceb86e67a-kube-api-access-5xcmx\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.929649 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.959629 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" event={"ID":"bd8ea216-286a-4575-8b3e-fd8c3f2afcda","Type":"ContainerDied","Data":"1df06d886a16aa120671056396b58c178ad7f09245735c77a0469e65125587e1"} Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.959998 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1df06d886a16aa120671056396b58c178ad7f09245735c77a0469e65125587e1" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.959649 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4" Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.962369 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qhvg" event={"ID":"d8f59b74-790e-44dd-828f-5fee749d88db","Type":"ContainerStarted","Data":"eb6b5491f25bb5576c4f054c8d2638c15b1f25c80576d4851400ccea4dd99b4a"} Mar 16 00:18:39 crc kubenswrapper[4751]: I0316 00:18:39.993485 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9qhvg" podStartSLOduration=1.563962061 podStartE2EDuration="3.993456865s" podCreationTimestamp="2026-03-16 00:18:36 +0000 UTC" firstStartedPulling="2026-03-16 00:18:36.927272169 +0000 UTC m=+745.196604449" lastFinishedPulling="2026-03-16 00:18:39.356766963 +0000 UTC m=+747.626099253" observedRunningTime="2026-03-16 00:18:39.990530017 +0000 UTC m=+748.259862307" watchObservedRunningTime="2026-03-16 00:18:39.993456865 +0000 UTC m=+748.262789195" Mar 16 00:18:40 crc kubenswrapper[4751]: I0316 00:18:40.148935 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm"] Mar 16 00:18:40 crc kubenswrapper[4751]: W0316 00:18:40.153420 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89cd9003_4d55_44e5_8f14_9caceb86e67a.slice/crio-c1261fa6da1f9ffe7fa2000118d6130b5abe686fedb2c35b903126bc2a95533e WatchSource:0}: Error finding container c1261fa6da1f9ffe7fa2000118d6130b5abe686fedb2c35b903126bc2a95533e: Status 404 returned error can't find the container with id c1261fa6da1f9ffe7fa2000118d6130b5abe686fedb2c35b903126bc2a95533e Mar 16 00:18:40 crc kubenswrapper[4751]: I0316 00:18:40.974434 4751 generic.go:334] "Generic (PLEG): container finished" podID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerID="6ea0c7aa6f69e4e6b5f3fa6a135cc9dfd5b2aae1e42a545c546acb9e08c0e4ef" exitCode=0 Mar 16 00:18:40 crc kubenswrapper[4751]: I0316 00:18:40.974523 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" event={"ID":"89cd9003-4d55-44e5-8f14-9caceb86e67a","Type":"ContainerDied","Data":"6ea0c7aa6f69e4e6b5f3fa6a135cc9dfd5b2aae1e42a545c546acb9e08c0e4ef"} Mar 16 00:18:40 crc kubenswrapper[4751]: I0316 00:18:40.975053 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" event={"ID":"89cd9003-4d55-44e5-8f14-9caceb86e67a","Type":"ContainerStarted","Data":"c1261fa6da1f9ffe7fa2000118d6130b5abe686fedb2c35b903126bc2a95533e"} Mar 16 00:18:43 crc kubenswrapper[4751]: I0316 00:18:43.993876 4751 generic.go:334] "Generic (PLEG): container finished" podID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerID="20f1a821accba78db9c67471f62355af4c15d2097c380d66ba4cb40a260b77c4" exitCode=0 Mar 16 00:18:43 crc kubenswrapper[4751]: I0316 00:18:43.993966 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" event={"ID":"89cd9003-4d55-44e5-8f14-9caceb86e67a","Type":"ContainerDied","Data":"20f1a821accba78db9c67471f62355af4c15d2097c380d66ba4cb40a260b77c4"} Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.165717 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v8jx7"] Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.166620 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.181087 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v8jx7"] Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.284833 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/e8f60066-6de6-43a8-8570-e67806d4d0dd-kube-api-access-slf9m\") pod \"certified-operators-v8jx7\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.284884 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-catalog-content\") pod \"certified-operators-v8jx7\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.284914 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-utilities\") pod \"certified-operators-v8jx7\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.385536 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/e8f60066-6de6-43a8-8570-e67806d4d0dd-kube-api-access-slf9m\") pod \"certified-operators-v8jx7\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.385840 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-catalog-content\") pod \"certified-operators-v8jx7\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.385955 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-utilities\") pod \"certified-operators-v8jx7\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.386287 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-catalog-content\") pod \"certified-operators-v8jx7\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.386433 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-utilities\") pod \"certified-operators-v8jx7\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.427149 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/e8f60066-6de6-43a8-8570-e67806d4d0dd-kube-api-access-slf9m\") pod \"certified-operators-v8jx7\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.479539 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:44 crc kubenswrapper[4751]: I0316 00:18:44.669819 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v8jx7"] Mar 16 00:18:44 crc kubenswrapper[4751]: W0316 00:18:44.672844 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8f60066_6de6_43a8_8570_e67806d4d0dd.slice/crio-1cffc92c9d5751517659969ae31b8ca2cc6e6a1e821a89378ea5a17eecf96a51 WatchSource:0}: Error finding container 1cffc92c9d5751517659969ae31b8ca2cc6e6a1e821a89378ea5a17eecf96a51: Status 404 returned error can't find the container with id 1cffc92c9d5751517659969ae31b8ca2cc6e6a1e821a89378ea5a17eecf96a51 Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:44.999975 4751 generic.go:334] "Generic (PLEG): container finished" podID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerID="62673708f304ca3c641eb86c5bee90b3be3473d1ec58c93d05ef70970944fe1e" exitCode=0 Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.000056 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" event={"ID":"89cd9003-4d55-44e5-8f14-9caceb86e67a","Type":"ContainerDied","Data":"62673708f304ca3c641eb86c5bee90b3be3473d1ec58c93d05ef70970944fe1e"} Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.001607 4751 generic.go:334] "Generic (PLEG): container finished" podID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerID="1b8239f6921fdabf14fc65462fcdf5a86d7f0b2a1f14006f7384672de15c041d" exitCode=0 Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.001632 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8jx7" event={"ID":"e8f60066-6de6-43a8-8570-e67806d4d0dd","Type":"ContainerDied","Data":"1b8239f6921fdabf14fc65462fcdf5a86d7f0b2a1f14006f7384672de15c041d"} Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.001648 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8jx7" event={"ID":"e8f60066-6de6-43a8-8570-e67806d4d0dd","Type":"ContainerStarted","Data":"1cffc92c9d5751517659969ae31b8ca2cc6e6a1e821a89378ea5a17eecf96a51"} Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.891045 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts"] Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.896212 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.913284 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts"] Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.918083 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.918193 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nhqn\" (UniqueName: \"kubernetes.io/projected/d54c0437-6747-476e-a69c-b7b59ee0bd7d-kube-api-access-8nhqn\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:45 crc kubenswrapper[4751]: I0316 00:18:45.918232 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.018999 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.019051 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nhqn\" (UniqueName: \"kubernetes.io/projected/d54c0437-6747-476e-a69c-b7b59ee0bd7d-kube-api-access-8nhqn\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.019087 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.019499 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.019700 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.046316 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nhqn\" (UniqueName: \"kubernetes.io/projected/d54c0437-6747-476e-a69c-b7b59ee0bd7d-kube-api-access-8nhqn\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.224936 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.300492 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.323687 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-bundle\") pod \"89cd9003-4d55-44e5-8f14-9caceb86e67a\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.323750 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-util\") pod \"89cd9003-4d55-44e5-8f14-9caceb86e67a\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.323858 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xcmx\" (UniqueName: \"kubernetes.io/projected/89cd9003-4d55-44e5-8f14-9caceb86e67a-kube-api-access-5xcmx\") pod \"89cd9003-4d55-44e5-8f14-9caceb86e67a\" (UID: \"89cd9003-4d55-44e5-8f14-9caceb86e67a\") " Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.324942 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-bundle" (OuterVolumeSpecName: "bundle") pod "89cd9003-4d55-44e5-8f14-9caceb86e67a" (UID: "89cd9003-4d55-44e5-8f14-9caceb86e67a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.332450 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89cd9003-4d55-44e5-8f14-9caceb86e67a-kube-api-access-5xcmx" (OuterVolumeSpecName: "kube-api-access-5xcmx") pod "89cd9003-4d55-44e5-8f14-9caceb86e67a" (UID: "89cd9003-4d55-44e5-8f14-9caceb86e67a"). InnerVolumeSpecName "kube-api-access-5xcmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.348183 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-util" (OuterVolumeSpecName: "util") pod "89cd9003-4d55-44e5-8f14-9caceb86e67a" (UID: "89cd9003-4d55-44e5-8f14-9caceb86e67a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.425769 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xcmx\" (UniqueName: \"kubernetes.io/projected/89cd9003-4d55-44e5-8f14-9caceb86e67a-kube-api-access-5xcmx\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.425814 4751 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.425827 4751 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89cd9003-4d55-44e5-8f14-9caceb86e67a-util\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.551380 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.552276 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:46 crc kubenswrapper[4751]: I0316 00:18:46.906621 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts"] Mar 16 00:18:47 crc kubenswrapper[4751]: I0316 00:18:47.023322 4751 generic.go:334] "Generic (PLEG): container finished" podID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerID="0f2235b95ce0b7d3d5a7aa29788063a1240522bf7beb920eaee1b0fc3781f2b4" exitCode=0 Mar 16 00:18:47 crc kubenswrapper[4751]: I0316 00:18:47.023382 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8jx7" event={"ID":"e8f60066-6de6-43a8-8570-e67806d4d0dd","Type":"ContainerDied","Data":"0f2235b95ce0b7d3d5a7aa29788063a1240522bf7beb920eaee1b0fc3781f2b4"} Mar 16 00:18:47 crc kubenswrapper[4751]: I0316 00:18:47.026059 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" event={"ID":"d54c0437-6747-476e-a69c-b7b59ee0bd7d","Type":"ContainerStarted","Data":"476813af2bb69a8ad148b08b29134d72241557fd03e8e1be441369dd0e6df344"} Mar 16 00:18:47 crc kubenswrapper[4751]: I0316 00:18:47.028698 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" Mar 16 00:18:47 crc kubenswrapper[4751]: I0316 00:18:47.029073 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm" event={"ID":"89cd9003-4d55-44e5-8f14-9caceb86e67a","Type":"ContainerDied","Data":"c1261fa6da1f9ffe7fa2000118d6130b5abe686fedb2c35b903126bc2a95533e"} Mar 16 00:18:47 crc kubenswrapper[4751]: I0316 00:18:47.029091 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1261fa6da1f9ffe7fa2000118d6130b5abe686fedb2c35b903126bc2a95533e" Mar 16 00:18:47 crc kubenswrapper[4751]: I0316 00:18:47.603016 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9qhvg" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" containerName="registry-server" probeResult="failure" output=< Mar 16 00:18:47 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Mar 16 00:18:47 crc kubenswrapper[4751]: > Mar 16 00:18:48 crc kubenswrapper[4751]: I0316 00:18:48.036549 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8jx7" event={"ID":"e8f60066-6de6-43a8-8570-e67806d4d0dd","Type":"ContainerStarted","Data":"208e0c411c0c3a93e8de8191224e96803bd5e8c7be2f4f749576dcb573e1d042"} Mar 16 00:18:48 crc kubenswrapper[4751]: I0316 00:18:48.038324 4751 generic.go:334] "Generic (PLEG): container finished" podID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerID="46339e2b9b2810651c70ad1afe4598ffdc3f0d579e71fcf8a8819cf4dc4520d0" exitCode=0 Mar 16 00:18:48 crc kubenswrapper[4751]: I0316 00:18:48.038349 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" event={"ID":"d54c0437-6747-476e-a69c-b7b59ee0bd7d","Type":"ContainerDied","Data":"46339e2b9b2810651c70ad1afe4598ffdc3f0d579e71fcf8a8819cf4dc4520d0"} Mar 16 00:18:48 crc kubenswrapper[4751]: I0316 00:18:48.094387 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v8jx7" podStartSLOduration=1.6881156389999998 podStartE2EDuration="4.094363795s" podCreationTimestamp="2026-03-16 00:18:44 +0000 UTC" firstStartedPulling="2026-03-16 00:18:45.002984674 +0000 UTC m=+753.272316964" lastFinishedPulling="2026-03-16 00:18:47.40923284 +0000 UTC m=+755.678565120" observedRunningTime="2026-03-16 00:18:48.081844419 +0000 UTC m=+756.351176709" watchObservedRunningTime="2026-03-16 00:18:48.094363795 +0000 UTC m=+756.363696085" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.153571 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t"] Mar 16 00:18:51 crc kubenswrapper[4751]: E0316 00:18:51.153772 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerName="extract" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.153784 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerName="extract" Mar 16 00:18:51 crc kubenswrapper[4751]: E0316 00:18:51.153795 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerName="util" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.153800 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerName="util" Mar 16 00:18:51 crc kubenswrapper[4751]: E0316 00:18:51.153809 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerName="pull" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.153815 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerName="pull" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.153896 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="89cd9003-4d55-44e5-8f14-9caceb86e67a" containerName="extract" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.154251 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.155781 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-bvdd7" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.156490 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.157801 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.167882 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t"] Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.272798 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv"] Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.274030 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.276478 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-qkps4" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.277075 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp"] Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.277489 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.278417 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.285347 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv"] Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.291038 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp"] Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.297989 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rn8b\" (UniqueName: \"kubernetes.io/projected/9a21e689-2895-407e-8888-b5e66e9f69d8-kube-api-access-2rn8b\") pod \"obo-prometheus-operator-68bc856cb9-5t56t\" (UID: \"9a21e689-2895-407e-8888-b5e66e9f69d8\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.387388 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-dkpf2"] Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.388182 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.391148 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-j82q5" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.391344 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.399243 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rn8b\" (UniqueName: \"kubernetes.io/projected/9a21e689-2895-407e-8888-b5e66e9f69d8-kube-api-access-2rn8b\") pod \"obo-prometheus-operator-68bc856cb9-5t56t\" (UID: \"9a21e689-2895-407e-8888-b5e66e9f69d8\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.399287 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/91271a6a-d14a-4681-a311-58ddc1253908-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp\" (UID: \"91271a6a-d14a-4681-a311-58ddc1253908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.399355 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8d263a57-8030-4502-8dab-5fcce4c4f85e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv\" (UID: \"8d263a57-8030-4502-8dab-5fcce4c4f85e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.399381 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/91271a6a-d14a-4681-a311-58ddc1253908-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp\" (UID: \"91271a6a-d14a-4681-a311-58ddc1253908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.399413 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8d263a57-8030-4502-8dab-5fcce4c4f85e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv\" (UID: \"8d263a57-8030-4502-8dab-5fcce4c4f85e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.404117 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-dkpf2"] Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.434441 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rn8b\" (UniqueName: \"kubernetes.io/projected/9a21e689-2895-407e-8888-b5e66e9f69d8-kube-api-access-2rn8b\") pod \"obo-prometheus-operator-68bc856cb9-5t56t\" (UID: \"9a21e689-2895-407e-8888-b5e66e9f69d8\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.474371 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.500237 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8d263a57-8030-4502-8dab-5fcce4c4f85e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv\" (UID: \"8d263a57-8030-4502-8dab-5fcce4c4f85e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.500308 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/91271a6a-d14a-4681-a311-58ddc1253908-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp\" (UID: \"91271a6a-d14a-4681-a311-58ddc1253908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.500374 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8d263a57-8030-4502-8dab-5fcce4c4f85e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv\" (UID: \"8d263a57-8030-4502-8dab-5fcce4c4f85e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.500403 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlxw2\" (UniqueName: \"kubernetes.io/projected/2469430b-1d52-419e-b6c7-74be1d58ba4b-kube-api-access-xlxw2\") pod \"observability-operator-59bdc8b94-dkpf2\" (UID: \"2469430b-1d52-419e-b6c7-74be1d58ba4b\") " pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.500423 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/91271a6a-d14a-4681-a311-58ddc1253908-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp\" (UID: \"91271a6a-d14a-4681-a311-58ddc1253908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.500443 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2469430b-1d52-419e-b6c7-74be1d58ba4b-observability-operator-tls\") pod \"observability-operator-59bdc8b94-dkpf2\" (UID: \"2469430b-1d52-419e-b6c7-74be1d58ba4b\") " pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.503743 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/91271a6a-d14a-4681-a311-58ddc1253908-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp\" (UID: \"91271a6a-d14a-4681-a311-58ddc1253908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.503762 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/91271a6a-d14a-4681-a311-58ddc1253908-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp\" (UID: \"91271a6a-d14a-4681-a311-58ddc1253908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.504584 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8d263a57-8030-4502-8dab-5fcce4c4f85e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv\" (UID: \"8d263a57-8030-4502-8dab-5fcce4c4f85e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.504689 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8d263a57-8030-4502-8dab-5fcce4c4f85e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv\" (UID: \"8d263a57-8030-4502-8dab-5fcce4c4f85e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.590090 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-8bc59"] Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.590833 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.593614 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-dgknk" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.601174 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2469430b-1d52-419e-b6c7-74be1d58ba4b-observability-operator-tls\") pod \"observability-operator-59bdc8b94-dkpf2\" (UID: \"2469430b-1d52-419e-b6c7-74be1d58ba4b\") " pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.601355 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlxw2\" (UniqueName: \"kubernetes.io/projected/2469430b-1d52-419e-b6c7-74be1d58ba4b-kube-api-access-xlxw2\") pod \"observability-operator-59bdc8b94-dkpf2\" (UID: \"2469430b-1d52-419e-b6c7-74be1d58ba4b\") " pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.601422 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-8bc59"] Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.605709 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.611691 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.619783 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2469430b-1d52-419e-b6c7-74be1d58ba4b-observability-operator-tls\") pod \"observability-operator-59bdc8b94-dkpf2\" (UID: \"2469430b-1d52-419e-b6c7-74be1d58ba4b\") " pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.630886 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlxw2\" (UniqueName: \"kubernetes.io/projected/2469430b-1d52-419e-b6c7-74be1d58ba4b-kube-api-access-xlxw2\") pod \"observability-operator-59bdc8b94-dkpf2\" (UID: \"2469430b-1d52-419e-b6c7-74be1d58ba4b\") " pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.702565 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8gwp\" (UniqueName: \"kubernetes.io/projected/c9b395e8-a7f9-4fee-b851-c4d382678e5c-kube-api-access-w8gwp\") pod \"perses-operator-5bf474d74f-8bc59\" (UID: \"c9b395e8-a7f9-4fee-b851-c4d382678e5c\") " pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.702656 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c9b395e8-a7f9-4fee-b851-c4d382678e5c-openshift-service-ca\") pod \"perses-operator-5bf474d74f-8bc59\" (UID: \"c9b395e8-a7f9-4fee-b851-c4d382678e5c\") " pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.705312 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.804402 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c9b395e8-a7f9-4fee-b851-c4d382678e5c-openshift-service-ca\") pod \"perses-operator-5bf474d74f-8bc59\" (UID: \"c9b395e8-a7f9-4fee-b851-c4d382678e5c\") " pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.804510 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8gwp\" (UniqueName: \"kubernetes.io/projected/c9b395e8-a7f9-4fee-b851-c4d382678e5c-kube-api-access-w8gwp\") pod \"perses-operator-5bf474d74f-8bc59\" (UID: \"c9b395e8-a7f9-4fee-b851-c4d382678e5c\") " pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.805360 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c9b395e8-a7f9-4fee-b851-c4d382678e5c-openshift-service-ca\") pod \"perses-operator-5bf474d74f-8bc59\" (UID: \"c9b395e8-a7f9-4fee-b851-c4d382678e5c\") " pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.823017 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8gwp\" (UniqueName: \"kubernetes.io/projected/c9b395e8-a7f9-4fee-b851-c4d382678e5c-kube-api-access-w8gwp\") pod \"perses-operator-5bf474d74f-8bc59\" (UID: \"c9b395e8-a7f9-4fee-b851-c4d382678e5c\") " pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:18:51 crc kubenswrapper[4751]: I0316 00:18:51.949072 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:18:52 crc kubenswrapper[4751]: W0316 00:18:52.536418 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91271a6a_d14a_4681_a311_58ddc1253908.slice/crio-9713d4b21277a6e17f4b40263a31ea190b17d597276464b4fd3f02378ffe49f9 WatchSource:0}: Error finding container 9713d4b21277a6e17f4b40263a31ea190b17d597276464b4fd3f02378ffe49f9: Status 404 returned error can't find the container with id 9713d4b21277a6e17f4b40263a31ea190b17d597276464b4fd3f02378ffe49f9 Mar 16 00:18:52 crc kubenswrapper[4751]: I0316 00:18:52.543422 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp"] Mar 16 00:18:52 crc kubenswrapper[4751]: I0316 00:18:52.582320 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv"] Mar 16 00:18:52 crc kubenswrapper[4751]: I0316 00:18:52.754297 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-8bc59"] Mar 16 00:18:52 crc kubenswrapper[4751]: I0316 00:18:52.761240 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t"] Mar 16 00:18:52 crc kubenswrapper[4751]: I0316 00:18:52.772689 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-dkpf2"] Mar 16 00:18:53 crc kubenswrapper[4751]: I0316 00:18:53.080617 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-8bc59" event={"ID":"c9b395e8-a7f9-4fee-b851-c4d382678e5c","Type":"ContainerStarted","Data":"e13833d50fdff74a833771e4d88cce1d7f19ad80f0e25e9447e22d6eac6d57e8"} Mar 16 00:18:53 crc kubenswrapper[4751]: I0316 00:18:53.082472 4751 generic.go:334] "Generic (PLEG): container finished" podID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerID="cf5aa8796256d96bc1db16e812bc1fa1c24e5ef5b19e71c6414587d85c62d3f5" exitCode=0 Mar 16 00:18:53 crc kubenswrapper[4751]: I0316 00:18:53.082509 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" event={"ID":"d54c0437-6747-476e-a69c-b7b59ee0bd7d","Type":"ContainerDied","Data":"cf5aa8796256d96bc1db16e812bc1fa1c24e5ef5b19e71c6414587d85c62d3f5"} Mar 16 00:18:53 crc kubenswrapper[4751]: I0316 00:18:53.084251 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" event={"ID":"91271a6a-d14a-4681-a311-58ddc1253908","Type":"ContainerStarted","Data":"9713d4b21277a6e17f4b40263a31ea190b17d597276464b4fd3f02378ffe49f9"} Mar 16 00:18:53 crc kubenswrapper[4751]: I0316 00:18:53.086228 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" event={"ID":"2469430b-1d52-419e-b6c7-74be1d58ba4b","Type":"ContainerStarted","Data":"cb6b0c736bc1a50aff50db3951127120c613f7bedd947fe492f4615723499ec3"} Mar 16 00:18:53 crc kubenswrapper[4751]: I0316 00:18:53.087194 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" event={"ID":"8d263a57-8030-4502-8dab-5fcce4c4f85e","Type":"ContainerStarted","Data":"256be407396046acb92ef16c428d9d5a3dd524f6fe1f6ae0d86bc7b9faef0c48"} Mar 16 00:18:53 crc kubenswrapper[4751]: I0316 00:18:53.088723 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t" event={"ID":"9a21e689-2895-407e-8888-b5e66e9f69d8","Type":"ContainerStarted","Data":"52b7c31f4ecb3f69b7613744ab2edd8a6e03b6d061dd8cbe084e2091fb3b3d09"} Mar 16 00:18:54 crc kubenswrapper[4751]: I0316 00:18:54.095619 4751 generic.go:334] "Generic (PLEG): container finished" podID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerID="92ece5ceb9bb3559b26155b24e057edd6b1b9365c83e464a09d4df7f70429e7b" exitCode=0 Mar 16 00:18:54 crc kubenswrapper[4751]: I0316 00:18:54.095664 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" event={"ID":"d54c0437-6747-476e-a69c-b7b59ee0bd7d","Type":"ContainerDied","Data":"92ece5ceb9bb3559b26155b24e057edd6b1b9365c83e464a09d4df7f70429e7b"} Mar 16 00:18:54 crc kubenswrapper[4751]: I0316 00:18:54.479629 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:54 crc kubenswrapper[4751]: I0316 00:18:54.479678 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:54 crc kubenswrapper[4751]: I0316 00:18:54.537500 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.219987 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.538197 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.593084 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nhqn\" (UniqueName: \"kubernetes.io/projected/d54c0437-6747-476e-a69c-b7b59ee0bd7d-kube-api-access-8nhqn\") pod \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.593145 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-bundle\") pod \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.593223 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-util\") pod \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\" (UID: \"d54c0437-6747-476e-a69c-b7b59ee0bd7d\") " Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.597655 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-bundle" (OuterVolumeSpecName: "bundle") pod "d54c0437-6747-476e-a69c-b7b59ee0bd7d" (UID: "d54c0437-6747-476e-a69c-b7b59ee0bd7d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.605368 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-util" (OuterVolumeSpecName: "util") pod "d54c0437-6747-476e-a69c-b7b59ee0bd7d" (UID: "d54c0437-6747-476e-a69c-b7b59ee0bd7d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.606271 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d54c0437-6747-476e-a69c-b7b59ee0bd7d-kube-api-access-8nhqn" (OuterVolumeSpecName: "kube-api-access-8nhqn") pod "d54c0437-6747-476e-a69c-b7b59ee0bd7d" (UID: "d54c0437-6747-476e-a69c-b7b59ee0bd7d"). InnerVolumeSpecName "kube-api-access-8nhqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.694814 4751 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-util\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.694849 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nhqn\" (UniqueName: \"kubernetes.io/projected/d54c0437-6747-476e-a69c-b7b59ee0bd7d-kube-api-access-8nhqn\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.694887 4751 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d54c0437-6747-476e-a69c-b7b59ee0bd7d-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.721587 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-59dd49d5d8-gpbf2"] Mar 16 00:18:55 crc kubenswrapper[4751]: E0316 00:18:55.721898 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerName="extract" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.721913 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerName="extract" Mar 16 00:18:55 crc kubenswrapper[4751]: E0316 00:18:55.721930 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerName="util" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.721935 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerName="util" Mar 16 00:18:55 crc kubenswrapper[4751]: E0316 00:18:55.721957 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerName="pull" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.721963 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerName="pull" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.722057 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d54c0437-6747-476e-a69c-b7b59ee0bd7d" containerName="extract" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.722459 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.725233 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.725568 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.725593 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-drjx9" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.725623 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.734920 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-59dd49d5d8-gpbf2"] Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.796395 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t8dw\" (UniqueName: \"kubernetes.io/projected/db4731c9-44cf-43e2-a3b5-19fb9511c6f7-kube-api-access-7t8dw\") pod \"elastic-operator-59dd49d5d8-gpbf2\" (UID: \"db4731c9-44cf-43e2-a3b5-19fb9511c6f7\") " pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.796458 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/db4731c9-44cf-43e2-a3b5-19fb9511c6f7-apiservice-cert\") pod \"elastic-operator-59dd49d5d8-gpbf2\" (UID: \"db4731c9-44cf-43e2-a3b5-19fb9511c6f7\") " pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.796647 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/db4731c9-44cf-43e2-a3b5-19fb9511c6f7-webhook-cert\") pod \"elastic-operator-59dd49d5d8-gpbf2\" (UID: \"db4731c9-44cf-43e2-a3b5-19fb9511c6f7\") " pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.898216 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/db4731c9-44cf-43e2-a3b5-19fb9511c6f7-apiservice-cert\") pod \"elastic-operator-59dd49d5d8-gpbf2\" (UID: \"db4731c9-44cf-43e2-a3b5-19fb9511c6f7\") " pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.898300 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/db4731c9-44cf-43e2-a3b5-19fb9511c6f7-webhook-cert\") pod \"elastic-operator-59dd49d5d8-gpbf2\" (UID: \"db4731c9-44cf-43e2-a3b5-19fb9511c6f7\") " pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.898338 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t8dw\" (UniqueName: \"kubernetes.io/projected/db4731c9-44cf-43e2-a3b5-19fb9511c6f7-kube-api-access-7t8dw\") pod \"elastic-operator-59dd49d5d8-gpbf2\" (UID: \"db4731c9-44cf-43e2-a3b5-19fb9511c6f7\") " pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.904948 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/db4731c9-44cf-43e2-a3b5-19fb9511c6f7-apiservice-cert\") pod \"elastic-operator-59dd49d5d8-gpbf2\" (UID: \"db4731c9-44cf-43e2-a3b5-19fb9511c6f7\") " pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.907890 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/db4731c9-44cf-43e2-a3b5-19fb9511c6f7-webhook-cert\") pod \"elastic-operator-59dd49d5d8-gpbf2\" (UID: \"db4731c9-44cf-43e2-a3b5-19fb9511c6f7\") " pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:55 crc kubenswrapper[4751]: I0316 00:18:55.933034 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t8dw\" (UniqueName: \"kubernetes.io/projected/db4731c9-44cf-43e2-a3b5-19fb9511c6f7-kube-api-access-7t8dw\") pod \"elastic-operator-59dd49d5d8-gpbf2\" (UID: \"db4731c9-44cf-43e2-a3b5-19fb9511c6f7\") " pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:56 crc kubenswrapper[4751]: I0316 00:18:56.052485 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" Mar 16 00:18:56 crc kubenswrapper[4751]: I0316 00:18:56.138941 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" event={"ID":"d54c0437-6747-476e-a69c-b7b59ee0bd7d","Type":"ContainerDied","Data":"476813af2bb69a8ad148b08b29134d72241557fd03e8e1be441369dd0e6df344"} Mar 16 00:18:56 crc kubenswrapper[4751]: I0316 00:18:56.138973 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts" Mar 16 00:18:56 crc kubenswrapper[4751]: I0316 00:18:56.139000 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="476813af2bb69a8ad148b08b29134d72241557fd03e8e1be441369dd0e6df344" Mar 16 00:18:56 crc kubenswrapper[4751]: I0316 00:18:56.402876 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-59dd49d5d8-gpbf2"] Mar 16 00:18:56 crc kubenswrapper[4751]: I0316 00:18:56.616404 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:56 crc kubenswrapper[4751]: I0316 00:18:56.757917 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:18:56 crc kubenswrapper[4751]: I0316 00:18:56.761806 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v8jx7"] Mar 16 00:18:57 crc kubenswrapper[4751]: I0316 00:18:57.144622 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v8jx7" podUID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerName="registry-server" containerID="cri-o://208e0c411c0c3a93e8de8191224e96803bd5e8c7be2f4f749576dcb573e1d042" gracePeriod=2 Mar 16 00:18:57 crc kubenswrapper[4751]: I0316 00:18:57.961906 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qhvg"] Mar 16 00:18:58 crc kubenswrapper[4751]: I0316 00:18:58.172245 4751 generic.go:334] "Generic (PLEG): container finished" podID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerID="208e0c411c0c3a93e8de8191224e96803bd5e8c7be2f4f749576dcb573e1d042" exitCode=0 Mar 16 00:18:58 crc kubenswrapper[4751]: I0316 00:18:58.172338 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8jx7" event={"ID":"e8f60066-6de6-43a8-8570-e67806d4d0dd","Type":"ContainerDied","Data":"208e0c411c0c3a93e8de8191224e96803bd5e8c7be2f4f749576dcb573e1d042"} Mar 16 00:18:58 crc kubenswrapper[4751]: I0316 00:18:58.172448 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9qhvg" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" containerName="registry-server" containerID="cri-o://eb6b5491f25bb5576c4f054c8d2638c15b1f25c80576d4851400ccea4dd99b4a" gracePeriod=2 Mar 16 00:18:59 crc kubenswrapper[4751]: I0316 00:18:59.184818 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" event={"ID":"db4731c9-44cf-43e2-a3b5-19fb9511c6f7","Type":"ContainerStarted","Data":"76e79e34aae908b67707526b5493fddd02b04b62bead0a5b50a037a018375387"} Mar 16 00:18:59 crc kubenswrapper[4751]: I0316 00:18:59.190224 4751 generic.go:334] "Generic (PLEG): container finished" podID="d8f59b74-790e-44dd-828f-5fee749d88db" containerID="eb6b5491f25bb5576c4f054c8d2638c15b1f25c80576d4851400ccea4dd99b4a" exitCode=0 Mar 16 00:18:59 crc kubenswrapper[4751]: I0316 00:18:59.190282 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qhvg" event={"ID":"d8f59b74-790e-44dd-828f-5fee749d88db","Type":"ContainerDied","Data":"eb6b5491f25bb5576c4f054c8d2638c15b1f25c80576d4851400ccea4dd99b4a"} Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.383204 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.508085 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh5rv\" (UniqueName: \"kubernetes.io/projected/d8f59b74-790e-44dd-828f-5fee749d88db-kube-api-access-vh5rv\") pod \"d8f59b74-790e-44dd-828f-5fee749d88db\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.508151 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-utilities\") pod \"d8f59b74-790e-44dd-828f-5fee749d88db\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.508218 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-catalog-content\") pod \"d8f59b74-790e-44dd-828f-5fee749d88db\" (UID: \"d8f59b74-790e-44dd-828f-5fee749d88db\") " Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.509066 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-utilities" (OuterVolumeSpecName: "utilities") pod "d8f59b74-790e-44dd-828f-5fee749d88db" (UID: "d8f59b74-790e-44dd-828f-5fee749d88db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.527906 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f59b74-790e-44dd-828f-5fee749d88db-kube-api-access-vh5rv" (OuterVolumeSpecName: "kube-api-access-vh5rv") pod "d8f59b74-790e-44dd-828f-5fee749d88db" (UID: "d8f59b74-790e-44dd-828f-5fee749d88db"). InnerVolumeSpecName "kube-api-access-vh5rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.609791 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh5rv\" (UniqueName: \"kubernetes.io/projected/d8f59b74-790e-44dd-828f-5fee749d88db-kube-api-access-vh5rv\") on node \"crc\" DevicePath \"\"" Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.609822 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.634409 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8f59b74-790e-44dd-828f-5fee749d88db" (UID: "d8f59b74-790e-44dd-828f-5fee749d88db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:19:01 crc kubenswrapper[4751]: I0316 00:19:01.711270 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f59b74-790e-44dd-828f-5fee749d88db-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.159679 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.212956 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qhvg" event={"ID":"d8f59b74-790e-44dd-828f-5fee749d88db","Type":"ContainerDied","Data":"58ac8595539517f1b415503d2714aaadb94ec31058dee58aba1c7744360c6f93"} Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.213011 4751 scope.go:117] "RemoveContainer" containerID="eb6b5491f25bb5576c4f054c8d2638c15b1f25c80576d4851400ccea4dd99b4a" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.213154 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qhvg" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.218310 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v8jx7" event={"ID":"e8f60066-6de6-43a8-8570-e67806d4d0dd","Type":"ContainerDied","Data":"1cffc92c9d5751517659969ae31b8ca2cc6e6a1e821a89378ea5a17eecf96a51"} Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.218399 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v8jx7" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.245432 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qhvg"] Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.249646 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9qhvg"] Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.318925 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-utilities\") pod \"e8f60066-6de6-43a8-8570-e67806d4d0dd\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.319298 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/e8f60066-6de6-43a8-8570-e67806d4d0dd-kube-api-access-slf9m\") pod \"e8f60066-6de6-43a8-8570-e67806d4d0dd\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.319317 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-catalog-content\") pod \"e8f60066-6de6-43a8-8570-e67806d4d0dd\" (UID: \"e8f60066-6de6-43a8-8570-e67806d4d0dd\") " Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.321262 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-utilities" (OuterVolumeSpecName: "utilities") pod "e8f60066-6de6-43a8-8570-e67806d4d0dd" (UID: "e8f60066-6de6-43a8-8570-e67806d4d0dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.330421 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8f60066-6de6-43a8-8570-e67806d4d0dd-kube-api-access-slf9m" (OuterVolumeSpecName: "kube-api-access-slf9m") pod "e8f60066-6de6-43a8-8570-e67806d4d0dd" (UID: "e8f60066-6de6-43a8-8570-e67806d4d0dd"). InnerVolumeSpecName "kube-api-access-slf9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.371225 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8f60066-6de6-43a8-8570-e67806d4d0dd" (UID: "e8f60066-6de6-43a8-8570-e67806d4d0dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.420147 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.420179 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/e8f60066-6de6-43a8-8570-e67806d4d0dd-kube-api-access-slf9m\") on node \"crc\" DevicePath \"\"" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.420188 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8f60066-6de6-43a8-8570-e67806d4d0dd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.462664 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" path="/var/lib/kubelet/pods/d8f59b74-790e-44dd-828f-5fee749d88db/volumes" Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.634051 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v8jx7"] Mar 16 00:19:02 crc kubenswrapper[4751]: I0316 00:19:02.638353 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v8jx7"] Mar 16 00:19:04 crc kubenswrapper[4751]: I0316 00:19:04.463033 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8f60066-6de6-43a8-8570-e67806d4d0dd" path="/var/lib/kubelet/pods/e8f60066-6de6-43a8-8570-e67806d4d0dd/volumes" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.180557 4751 scope.go:117] "RemoveContainer" containerID="31dd9a6f4f9d2c699906be7b1a74aebb8679e2bb4a3ea1437049eb86a56c7632" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.220936 4751 scope.go:117] "RemoveContainer" containerID="87cb230e571d82f28acff8dfcda133d69e99b577a91f5de28596f5e60e542f94" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.248194 4751 scope.go:117] "RemoveContainer" containerID="208e0c411c0c3a93e8de8191224e96803bd5e8c7be2f4f749576dcb573e1d042" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.289571 4751 scope.go:117] "RemoveContainer" containerID="0f2235b95ce0b7d3d5a7aa29788063a1240522bf7beb920eaee1b0fc3781f2b4" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.327973 4751 scope.go:117] "RemoveContainer" containerID="1b8239f6921fdabf14fc65462fcdf5a86d7f0b2a1f14006f7384672de15c041d" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.943929 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f"] Mar 16 00:19:06 crc kubenswrapper[4751]: E0316 00:19:06.944371 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerName="extract-utilities" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.944387 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerName="extract-utilities" Mar 16 00:19:06 crc kubenswrapper[4751]: E0316 00:19:06.944396 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" containerName="extract-content" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.944403 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" containerName="extract-content" Mar 16 00:19:06 crc kubenswrapper[4751]: E0316 00:19:06.944412 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" containerName="extract-utilities" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.944419 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" containerName="extract-utilities" Mar 16 00:19:06 crc kubenswrapper[4751]: E0316 00:19:06.944430 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerName="extract-content" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.944436 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerName="extract-content" Mar 16 00:19:06 crc kubenswrapper[4751]: E0316 00:19:06.944442 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" containerName="registry-server" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.944448 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" containerName="registry-server" Mar 16 00:19:06 crc kubenswrapper[4751]: E0316 00:19:06.944463 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerName="registry-server" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.944469 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerName="registry-server" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.944561 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8f60066-6de6-43a8-8570-e67806d4d0dd" containerName="registry-server" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.944573 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f59b74-790e-44dd-828f-5fee749d88db" containerName="registry-server" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.944944 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.946465 4751 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-8k5kg" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.946651 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.950168 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Mar 16 00:19:06 crc kubenswrapper[4751]: I0316 00:19:06.966215 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f"] Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.135234 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4vn8\" (UniqueName: \"kubernetes.io/projected/86684463-85bd-47ae-87ba-38474ffac4d7-kube-api-access-s4vn8\") pod \"cert-manager-operator-controller-manager-5586865c96-vzz2f\" (UID: \"86684463-85bd-47ae-87ba-38474ffac4d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.135507 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/86684463-85bd-47ae-87ba-38474ffac4d7-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-vzz2f\" (UID: \"86684463-85bd-47ae-87ba-38474ffac4d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.236395 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/86684463-85bd-47ae-87ba-38474ffac4d7-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-vzz2f\" (UID: \"86684463-85bd-47ae-87ba-38474ffac4d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.236504 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4vn8\" (UniqueName: \"kubernetes.io/projected/86684463-85bd-47ae-87ba-38474ffac4d7-kube-api-access-s4vn8\") pod \"cert-manager-operator-controller-manager-5586865c96-vzz2f\" (UID: \"86684463-85bd-47ae-87ba-38474ffac4d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.236989 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/86684463-85bd-47ae-87ba-38474ffac4d7-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-vzz2f\" (UID: \"86684463-85bd-47ae-87ba-38474ffac4d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.261909 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" event={"ID":"2469430b-1d52-419e-b6c7-74be1d58ba4b","Type":"ContainerStarted","Data":"1d90ac02e6ec48765fcde5fc60e8fc0a99ffde97343ec13be550664a822cb844"} Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.263535 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.265821 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4vn8\" (UniqueName: \"kubernetes.io/projected/86684463-85bd-47ae-87ba-38474ffac4d7-kube-api-access-s4vn8\") pod \"cert-manager-operator-controller-manager-5586865c96-vzz2f\" (UID: \"86684463-85bd-47ae-87ba-38474ffac4d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.290435 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.290701 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" event={"ID":"8d263a57-8030-4502-8dab-5fcce4c4f85e","Type":"ContainerStarted","Data":"83a79c54d5d12b582bc0fdab99e83004aa7ddd5518670675e7921140e1670e09"} Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.292611 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t" event={"ID":"9a21e689-2895-407e-8888-b5e66e9f69d8","Type":"ContainerStarted","Data":"098eaafc620dc09bff497e203d523ef011a8ebc00da52cbffeaf5c6cb6e80b87"} Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.296805 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-8bc59" event={"ID":"c9b395e8-a7f9-4fee-b851-c4d382678e5c","Type":"ContainerStarted","Data":"aab125891032730c71c52fe1af89bde31a4404436f7909827248f3f804866aa6"} Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.297215 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.318356 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" event={"ID":"91271a6a-d14a-4681-a311-58ddc1253908","Type":"ContainerStarted","Data":"592490a2d7b737e2bd9c7fb5edda1784e4035f76c046b9f4d77d25bf96091a04"} Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.320064 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" event={"ID":"db4731c9-44cf-43e2-a3b5-19fb9511c6f7","Type":"ContainerStarted","Data":"ad828ba6b1d233e01a662532152f2918617ac553d426d75b53f905b06d31f90f"} Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.344165 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-dkpf2" podStartSLOduration=2.887930415 podStartE2EDuration="16.344148684s" podCreationTimestamp="2026-03-16 00:18:51 +0000 UTC" firstStartedPulling="2026-03-16 00:18:52.766338045 +0000 UTC m=+761.035670325" lastFinishedPulling="2026-03-16 00:19:06.222556294 +0000 UTC m=+774.491888594" observedRunningTime="2026-03-16 00:19:07.317915988 +0000 UTC m=+775.587248398" watchObservedRunningTime="2026-03-16 00:19:07.344148684 +0000 UTC m=+775.613480974" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.359568 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-5t56t" podStartSLOduration=2.946386065 podStartE2EDuration="16.359555578s" podCreationTimestamp="2026-03-16 00:18:51 +0000 UTC" firstStartedPulling="2026-03-16 00:18:52.767400956 +0000 UTC m=+761.036733246" lastFinishedPulling="2026-03-16 00:19:06.180570429 +0000 UTC m=+774.449902759" observedRunningTime="2026-03-16 00:19:07.357615499 +0000 UTC m=+775.626947789" watchObservedRunningTime="2026-03-16 00:19:07.359555578 +0000 UTC m=+775.628887868" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.383711 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv" podStartSLOduration=2.75701382 podStartE2EDuration="16.383698542s" podCreationTimestamp="2026-03-16 00:18:51 +0000 UTC" firstStartedPulling="2026-03-16 00:18:52.622273972 +0000 UTC m=+760.891606252" lastFinishedPulling="2026-03-16 00:19:06.248958674 +0000 UTC m=+774.518290974" observedRunningTime="2026-03-16 00:19:07.380084541 +0000 UTC m=+775.649416831" watchObservedRunningTime="2026-03-16 00:19:07.383698542 +0000 UTC m=+775.653030832" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.412019 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-8bc59" podStartSLOduration=2.95529788 podStartE2EDuration="16.412001769s" podCreationTimestamp="2026-03-16 00:18:51 +0000 UTC" firstStartedPulling="2026-03-16 00:18:52.768304194 +0000 UTC m=+761.037636484" lastFinishedPulling="2026-03-16 00:19:06.225008073 +0000 UTC m=+774.494340373" observedRunningTime="2026-03-16 00:19:07.406576422 +0000 UTC m=+775.675908712" watchObservedRunningTime="2026-03-16 00:19:07.412001769 +0000 UTC m=+775.681334059" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.431375 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp" podStartSLOduration=3.600567492 podStartE2EDuration="16.43135422s" podCreationTimestamp="2026-03-16 00:18:51 +0000 UTC" firstStartedPulling="2026-03-16 00:18:52.537832011 +0000 UTC m=+760.807164301" lastFinishedPulling="2026-03-16 00:19:05.368618739 +0000 UTC m=+773.637951029" observedRunningTime="2026-03-16 00:19:07.425957254 +0000 UTC m=+775.695289544" watchObservedRunningTime="2026-03-16 00:19:07.43135422 +0000 UTC m=+775.700686510" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.457177 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-59dd49d5d8-gpbf2" podStartSLOduration=5.046215277 podStartE2EDuration="12.457098546s" podCreationTimestamp="2026-03-16 00:18:55 +0000 UTC" firstStartedPulling="2026-03-16 00:18:58.814586623 +0000 UTC m=+767.083918903" lastFinishedPulling="2026-03-16 00:19:06.225469882 +0000 UTC m=+774.494802172" observedRunningTime="2026-03-16 00:19:07.45476721 +0000 UTC m=+775.724099520" watchObservedRunningTime="2026-03-16 00:19:07.457098546 +0000 UTC m=+775.726430866" Mar 16 00:19:07 crc kubenswrapper[4751]: I0316 00:19:07.556867 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.024640 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f"] Mar 16 00:19:08 crc kubenswrapper[4751]: W0316 00:19:08.025352 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86684463_85bd_47ae_87ba_38474ffac4d7.slice/crio-2012dadea8bbddccf2174504a230deea7751c46b8b0380f857fd6a16d3a6c569 WatchSource:0}: Error finding container 2012dadea8bbddccf2174504a230deea7751c46b8b0380f857fd6a16d3a6c569: Status 404 returned error can't find the container with id 2012dadea8bbddccf2174504a230deea7751c46b8b0380f857fd6a16d3a6c569 Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.027397 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.326066 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" event={"ID":"86684463-85bd-47ae-87ba-38474ffac4d7","Type":"ContainerStarted","Data":"2012dadea8bbddccf2174504a230deea7751c46b8b0380f857fd6a16d3a6c569"} Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.648928 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.650063 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.651915 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.652479 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.652954 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-bjk6k" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.652988 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.653090 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.653157 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.653304 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.654672 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.654868 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.674413 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760303 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760356 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760388 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760406 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760429 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/f222ef67-630e-4203-8857-b2d4fb4f1e89-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760753 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760806 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760842 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760877 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.760978 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.761023 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.761089 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.761158 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.761204 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.761264 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.861964 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862019 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862044 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862060 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862089 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862154 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862179 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862206 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862236 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862266 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862290 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862323 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862354 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862377 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862399 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/f222ef67-630e-4203-8857-b2d4fb4f1e89-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862743 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.862806 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.863375 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.863449 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.863660 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.863772 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.863808 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/f222ef67-630e-4203-8857-b2d4fb4f1e89-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.864286 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.867687 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.867843 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.868052 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/f222ef67-630e-4203-8857-b2d4fb4f1e89-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.871597 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.871605 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.871786 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.876010 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/f222ef67-630e-4203-8857-b2d4fb4f1e89-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"f222ef67-630e-4203-8857-b2d4fb4f1e89\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:08 crc kubenswrapper[4751]: I0316 00:19:08.969607 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:09 crc kubenswrapper[4751]: I0316 00:19:09.302138 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 16 00:19:09 crc kubenswrapper[4751]: I0316 00:19:09.340704 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"f222ef67-630e-4203-8857-b2d4fb4f1e89","Type":"ContainerStarted","Data":"bc36cc75339f6fa236c5794bc04b022750ea750ad66d6e1306a9dd0025d86df8"} Mar 16 00:19:11 crc kubenswrapper[4751]: I0316 00:19:11.952462 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-8bc59" Mar 16 00:19:12 crc kubenswrapper[4751]: I0316 00:19:12.364046 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" event={"ID":"86684463-85bd-47ae-87ba-38474ffac4d7","Type":"ContainerStarted","Data":"32d046a77253eceddcaa0189eef9591d3bc671628de097b04aced51edf2874e1"} Mar 16 00:19:12 crc kubenswrapper[4751]: I0316 00:19:12.397779 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-vzz2f" podStartSLOduration=3.107545551 podStartE2EDuration="6.397757064s" podCreationTimestamp="2026-03-16 00:19:06 +0000 UTC" firstStartedPulling="2026-03-16 00:19:08.027159948 +0000 UTC m=+776.296492238" lastFinishedPulling="2026-03-16 00:19:11.317371461 +0000 UTC m=+779.586703751" observedRunningTime="2026-03-16 00:19:12.397533038 +0000 UTC m=+780.666865338" watchObservedRunningTime="2026-03-16 00:19:12.397757064 +0000 UTC m=+780.667089364" Mar 16 00:19:15 crc kubenswrapper[4751]: I0316 00:19:15.872569 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-q9rzm"] Mar 16 00:19:15 crc kubenswrapper[4751]: I0316 00:19:15.873467 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:15 crc kubenswrapper[4751]: I0316 00:19:15.875077 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 16 00:19:15 crc kubenswrapper[4751]: I0316 00:19:15.875100 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 16 00:19:15 crc kubenswrapper[4751]: I0316 00:19:15.875118 4751 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-2vvx2" Mar 16 00:19:15 crc kubenswrapper[4751]: I0316 00:19:15.881669 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-q9rzm"] Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.060668 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb5cd\" (UniqueName: \"kubernetes.io/projected/f9b31035-18b9-4a25-b230-7f448975ea02-kube-api-access-bb5cd\") pod \"cert-manager-webhook-6888856db4-q9rzm\" (UID: \"f9b31035-18b9-4a25-b230-7f448975ea02\") " pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.060708 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f9b31035-18b9-4a25-b230-7f448975ea02-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-q9rzm\" (UID: \"f9b31035-18b9-4a25-b230-7f448975ea02\") " pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.161545 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb5cd\" (UniqueName: \"kubernetes.io/projected/f9b31035-18b9-4a25-b230-7f448975ea02-kube-api-access-bb5cd\") pod \"cert-manager-webhook-6888856db4-q9rzm\" (UID: \"f9b31035-18b9-4a25-b230-7f448975ea02\") " pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.161590 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f9b31035-18b9-4a25-b230-7f448975ea02-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-q9rzm\" (UID: \"f9b31035-18b9-4a25-b230-7f448975ea02\") " pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.182819 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f9b31035-18b9-4a25-b230-7f448975ea02-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-q9rzm\" (UID: \"f9b31035-18b9-4a25-b230-7f448975ea02\") " pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.187882 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb5cd\" (UniqueName: \"kubernetes.io/projected/f9b31035-18b9-4a25-b230-7f448975ea02-kube-api-access-bb5cd\") pod \"cert-manager-webhook-6888856db4-q9rzm\" (UID: \"f9b31035-18b9-4a25-b230-7f448975ea02\") " pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.188555 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.371592 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-wc58h"] Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.372263 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" Mar 16 00:19:16 crc kubenswrapper[4751]: W0316 00:19:16.374063 4751 reflector.go:561] object-"cert-manager"/"cert-manager-cainjector-dockercfg-qmpd7": failed to list *v1.Secret: secrets "cert-manager-cainjector-dockercfg-qmpd7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Mar 16 00:19:16 crc kubenswrapper[4751]: E0316 00:19:16.374129 4751 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-qmpd7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-manager-cainjector-dockercfg-qmpd7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.406539 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-wc58h"] Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.466275 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxf58\" (UniqueName: \"kubernetes.io/projected/c3bc081d-6e79-44f0-bc4e-e3c30474aa2c-kube-api-access-nxf58\") pod \"cert-manager-cainjector-5545bd876-wc58h\" (UID: \"c3bc081d-6e79-44f0-bc4e-e3c30474aa2c\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.466337 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3bc081d-6e79-44f0-bc4e-e3c30474aa2c-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-wc58h\" (UID: \"c3bc081d-6e79-44f0-bc4e-e3c30474aa2c\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.567819 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3bc081d-6e79-44f0-bc4e-e3c30474aa2c-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-wc58h\" (UID: \"c3bc081d-6e79-44f0-bc4e-e3c30474aa2c\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.567922 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxf58\" (UniqueName: \"kubernetes.io/projected/c3bc081d-6e79-44f0-bc4e-e3c30474aa2c-kube-api-access-nxf58\") pod \"cert-manager-cainjector-5545bd876-wc58h\" (UID: \"c3bc081d-6e79-44f0-bc4e-e3c30474aa2c\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.590244 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxf58\" (UniqueName: \"kubernetes.io/projected/c3bc081d-6e79-44f0-bc4e-e3c30474aa2c-kube-api-access-nxf58\") pod \"cert-manager-cainjector-5545bd876-wc58h\" (UID: \"c3bc081d-6e79-44f0-bc4e-e3c30474aa2c\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" Mar 16 00:19:16 crc kubenswrapper[4751]: I0316 00:19:16.590634 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3bc081d-6e79-44f0-bc4e-e3c30474aa2c-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-wc58h\" (UID: \"c3bc081d-6e79-44f0-bc4e-e3c30474aa2c\") " pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" Mar 16 00:19:17 crc kubenswrapper[4751]: I0316 00:19:17.541624 4751 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-qmpd7" Mar 16 00:19:17 crc kubenswrapper[4751]: I0316 00:19:17.547189 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" Mar 16 00:19:18 crc kubenswrapper[4751]: I0316 00:19:18.612983 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-q9rzm"] Mar 16 00:19:18 crc kubenswrapper[4751]: I0316 00:19:18.727785 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-wc58h"] Mar 16 00:19:18 crc kubenswrapper[4751]: W0316 00:19:18.733762 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3bc081d_6e79_44f0_bc4e_e3c30474aa2c.slice/crio-69af01b1c49a0ec536840162fce5df72913c2d9a666f1a9be80410acf28f0755 WatchSource:0}: Error finding container 69af01b1c49a0ec536840162fce5df72913c2d9a666f1a9be80410acf28f0755: Status 404 returned error can't find the container with id 69af01b1c49a0ec536840162fce5df72913c2d9a666f1a9be80410acf28f0755 Mar 16 00:19:19 crc kubenswrapper[4751]: I0316 00:19:19.456989 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" event={"ID":"c3bc081d-6e79-44f0-bc4e-e3c30474aa2c","Type":"ContainerStarted","Data":"69af01b1c49a0ec536840162fce5df72913c2d9a666f1a9be80410acf28f0755"} Mar 16 00:19:19 crc kubenswrapper[4751]: I0316 00:19:19.457730 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" event={"ID":"f9b31035-18b9-4a25-b230-7f448975ea02","Type":"ContainerStarted","Data":"9a23343be984ad3a13c2f7ae277497a94850ed3ee6fd0eddd7bf99cb0ab165ea"} Mar 16 00:19:29 crc kubenswrapper[4751]: I0316 00:19:29.527020 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"f222ef67-630e-4203-8857-b2d4fb4f1e89","Type":"ContainerStarted","Data":"5b82b50d41b7e69b6c2f05e068eec7d7cfe35d571897d03abb5f25efb577f443"} Mar 16 00:19:29 crc kubenswrapper[4751]: I0316 00:19:29.529330 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" event={"ID":"f9b31035-18b9-4a25-b230-7f448975ea02","Type":"ContainerStarted","Data":"fefa9fed93c37e889a3a93ad07ca5a644d25462b152314d55f9b421f5b1dde3c"} Mar 16 00:19:29 crc kubenswrapper[4751]: I0316 00:19:29.529450 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:29 crc kubenswrapper[4751]: I0316 00:19:29.531215 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" event={"ID":"c3bc081d-6e79-44f0-bc4e-e3c30474aa2c","Type":"ContainerStarted","Data":"dd905b758510a9448ef1712ee8b89d2a93f6c883b389d8075aff2a42f4283239"} Mar 16 00:19:29 crc kubenswrapper[4751]: I0316 00:19:29.600985 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" podStartSLOduration=4.418218571 podStartE2EDuration="14.600958387s" podCreationTimestamp="2026-03-16 00:19:15 +0000 UTC" firstStartedPulling="2026-03-16 00:19:18.632068333 +0000 UTC m=+786.901400623" lastFinishedPulling="2026-03-16 00:19:28.814808139 +0000 UTC m=+797.084140439" observedRunningTime="2026-03-16 00:19:29.5961153 +0000 UTC m=+797.865447590" watchObservedRunningTime="2026-03-16 00:19:29.600958387 +0000 UTC m=+797.870290687" Mar 16 00:19:29 crc kubenswrapper[4751]: I0316 00:19:29.617062 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-wc58h" podStartSLOduration=3.463280129 podStartE2EDuration="13.616952647s" podCreationTimestamp="2026-03-16 00:19:16 +0000 UTC" firstStartedPulling="2026-03-16 00:19:18.736208604 +0000 UTC m=+787.005540894" lastFinishedPulling="2026-03-16 00:19:28.889881112 +0000 UTC m=+797.159213412" observedRunningTime="2026-03-16 00:19:29.614125587 +0000 UTC m=+797.883457877" watchObservedRunningTime="2026-03-16 00:19:29.616952647 +0000 UTC m=+797.886284937" Mar 16 00:19:29 crc kubenswrapper[4751]: I0316 00:19:29.684413 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 16 00:19:29 crc kubenswrapper[4751]: I0316 00:19:29.738845 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 16 00:19:31 crc kubenswrapper[4751]: I0316 00:19:31.547944 4751 generic.go:334] "Generic (PLEG): container finished" podID="f222ef67-630e-4203-8857-b2d4fb4f1e89" containerID="5b82b50d41b7e69b6c2f05e068eec7d7cfe35d571897d03abb5f25efb577f443" exitCode=0 Mar 16 00:19:31 crc kubenswrapper[4751]: I0316 00:19:31.548014 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"f222ef67-630e-4203-8857-b2d4fb4f1e89","Type":"ContainerDied","Data":"5b82b50d41b7e69b6c2f05e068eec7d7cfe35d571897d03abb5f25efb577f443"} Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.187342 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.189647 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.192279 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-global-ca" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.192838 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-sys-config" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.193599 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-2bxtc" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.193846 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-framework-index-dockercfg" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.194949 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-ca" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.197874 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.358770 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwq7w\" (UniqueName: \"kubernetes.io/projected/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-kube-api-access-fwq7w\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.358828 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.358856 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.358875 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.358896 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.358913 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.358927 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-2bxtc-push\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.358948 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-2bxtc-pull\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.359087 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.359121 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.359146 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.359283 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.359345 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460507 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460558 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-2bxtc-push\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460594 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-2bxtc-pull\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460625 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460663 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460704 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460748 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460777 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460817 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwq7w\" (UniqueName: \"kubernetes.io/projected/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-kube-api-access-fwq7w\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.461020 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.461422 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.461594 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.461722 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.461976 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.460857 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.462335 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.463315 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.463377 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.463425 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.464269 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.464595 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.465296 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.470171 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-2bxtc-pull\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.470192 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.470303 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-2bxtc-push\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.482415 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwq7w\" (UniqueName: \"kubernetes.io/projected/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-kube-api-access-fwq7w\") pod \"service-telemetry-framework-index-1-build\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.567811 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.864130 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-rv7zp"] Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.866089 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-rv7zp" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.869188 4751 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-xqxw6" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.871225 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-rv7zp"] Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.970411 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pknn\" (UniqueName: \"kubernetes.io/projected/5c3f99b3-1665-4852-b3ab-e7b44e5593a4-kube-api-access-6pknn\") pod \"cert-manager-545d4d4674-rv7zp\" (UID: \"5c3f99b3-1665-4852-b3ab-e7b44e5593a4\") " pod="cert-manager/cert-manager-545d4d4674-rv7zp" Mar 16 00:19:34 crc kubenswrapper[4751]: I0316 00:19:34.970550 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5c3f99b3-1665-4852-b3ab-e7b44e5593a4-bound-sa-token\") pod \"cert-manager-545d4d4674-rv7zp\" (UID: \"5c3f99b3-1665-4852-b3ab-e7b44e5593a4\") " pod="cert-manager/cert-manager-545d4d4674-rv7zp" Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.069939 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.072711 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pknn\" (UniqueName: \"kubernetes.io/projected/5c3f99b3-1665-4852-b3ab-e7b44e5593a4-kube-api-access-6pknn\") pod \"cert-manager-545d4d4674-rv7zp\" (UID: \"5c3f99b3-1665-4852-b3ab-e7b44e5593a4\") " pod="cert-manager/cert-manager-545d4d4674-rv7zp" Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.072808 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5c3f99b3-1665-4852-b3ab-e7b44e5593a4-bound-sa-token\") pod \"cert-manager-545d4d4674-rv7zp\" (UID: \"5c3f99b3-1665-4852-b3ab-e7b44e5593a4\") " pod="cert-manager/cert-manager-545d4d4674-rv7zp" Mar 16 00:19:35 crc kubenswrapper[4751]: W0316 00:19:35.086151 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40097d75_fb3f_4891_8f6f_f90d72ee8ca0.slice/crio-0047df3baab30eea47b72c3c73032018e0943a56a133f3e81d50a38c09fee600 WatchSource:0}: Error finding container 0047df3baab30eea47b72c3c73032018e0943a56a133f3e81d50a38c09fee600: Status 404 returned error can't find the container with id 0047df3baab30eea47b72c3c73032018e0943a56a133f3e81d50a38c09fee600 Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.104933 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pknn\" (UniqueName: \"kubernetes.io/projected/5c3f99b3-1665-4852-b3ab-e7b44e5593a4-kube-api-access-6pknn\") pod \"cert-manager-545d4d4674-rv7zp\" (UID: \"5c3f99b3-1665-4852-b3ab-e7b44e5593a4\") " pod="cert-manager/cert-manager-545d4d4674-rv7zp" Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.105442 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5c3f99b3-1665-4852-b3ab-e7b44e5593a4-bound-sa-token\") pod \"cert-manager-545d4d4674-rv7zp\" (UID: \"5c3f99b3-1665-4852-b3ab-e7b44e5593a4\") " pod="cert-manager/cert-manager-545d4d4674-rv7zp" Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.187204 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-rv7zp" Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.477921 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-rv7zp"] Mar 16 00:19:35 crc kubenswrapper[4751]: W0316 00:19:35.486122 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c3f99b3_1665_4852_b3ab_e7b44e5593a4.slice/crio-17654e9fae634ef4570b49ce2ea766f269894e9c770001795dc6b7ae51ae9b9d WatchSource:0}: Error finding container 17654e9fae634ef4570b49ce2ea766f269894e9c770001795dc6b7ae51ae9b9d: Status 404 returned error can't find the container with id 17654e9fae634ef4570b49ce2ea766f269894e9c770001795dc6b7ae51ae9b9d Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.586878 4751 generic.go:334] "Generic (PLEG): container finished" podID="f222ef67-630e-4203-8857-b2d4fb4f1e89" containerID="e938485b714aa92b1b3991f68125f6d87788270c7822eb6def26f868f7ea62d2" exitCode=0 Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.586982 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"f222ef67-630e-4203-8857-b2d4fb4f1e89","Type":"ContainerDied","Data":"e938485b714aa92b1b3991f68125f6d87788270c7822eb6def26f868f7ea62d2"} Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.589180 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-rv7zp" event={"ID":"5c3f99b3-1665-4852-b3ab-e7b44e5593a4","Type":"ContainerStarted","Data":"17654e9fae634ef4570b49ce2ea766f269894e9c770001795dc6b7ae51ae9b9d"} Mar 16 00:19:35 crc kubenswrapper[4751]: I0316 00:19:35.592778 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"40097d75-fb3f-4891-8f6f-f90d72ee8ca0","Type":"ContainerStarted","Data":"0047df3baab30eea47b72c3c73032018e0943a56a133f3e81d50a38c09fee600"} Mar 16 00:19:36 crc kubenswrapper[4751]: I0316 00:19:36.192024 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-q9rzm" Mar 16 00:19:36 crc kubenswrapper[4751]: I0316 00:19:36.600554 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"f222ef67-630e-4203-8857-b2d4fb4f1e89","Type":"ContainerStarted","Data":"5eef55cb9ab935d5ad811023305f8720d11553363463f2a0b90b99e7456e767b"} Mar 16 00:19:36 crc kubenswrapper[4751]: I0316 00:19:36.600835 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:19:36 crc kubenswrapper[4751]: I0316 00:19:36.603227 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-rv7zp" event={"ID":"5c3f99b3-1665-4852-b3ab-e7b44e5593a4","Type":"ContainerStarted","Data":"71119e909eb22080f502fbd744b1b1cf21ae79b326e8c47a56cd707a4e59e3b0"} Mar 16 00:19:36 crc kubenswrapper[4751]: I0316 00:19:36.632550 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=8.912532004 podStartE2EDuration="28.632535327s" podCreationTimestamp="2026-03-16 00:19:08 +0000 UTC" firstStartedPulling="2026-03-16 00:19:09.334227915 +0000 UTC m=+777.603560205" lastFinishedPulling="2026-03-16 00:19:29.054231238 +0000 UTC m=+797.323563528" observedRunningTime="2026-03-16 00:19:36.632153736 +0000 UTC m=+804.901486046" watchObservedRunningTime="2026-03-16 00:19:36.632535327 +0000 UTC m=+804.901867617" Mar 16 00:19:36 crc kubenswrapper[4751]: I0316 00:19:36.649561 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-rv7zp" podStartSLOduration=2.649537875 podStartE2EDuration="2.649537875s" podCreationTimestamp="2026-03-16 00:19:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:19:36.646461929 +0000 UTC m=+804.915794229" watchObservedRunningTime="2026-03-16 00:19:36.649537875 +0000 UTC m=+804.918870165" Mar 16 00:19:42 crc kubenswrapper[4751]: I0316 00:19:42.659661 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"40097d75-fb3f-4891-8f6f-f90d72ee8ca0","Type":"ContainerStarted","Data":"231041d8616c6271d04a114bfc2883840b793cfb5c298ec823c85de6abc1a1e7"} Mar 16 00:19:43 crc kubenswrapper[4751]: I0316 00:19:43.667682 4751 generic.go:334] "Generic (PLEG): container finished" podID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerID="231041d8616c6271d04a114bfc2883840b793cfb5c298ec823c85de6abc1a1e7" exitCode=0 Mar 16 00:19:43 crc kubenswrapper[4751]: I0316 00:19:43.667736 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"40097d75-fb3f-4891-8f6f-f90d72ee8ca0","Type":"ContainerDied","Data":"231041d8616c6271d04a114bfc2883840b793cfb5c298ec823c85de6abc1a1e7"} Mar 16 00:19:44 crc kubenswrapper[4751]: I0316 00:19:44.677892 4751 generic.go:334] "Generic (PLEG): container finished" podID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerID="1a4c6d5f84b581922957d9ffe129a1b197b6a75509a592c116d7a9dfb1599faf" exitCode=0 Mar 16 00:19:44 crc kubenswrapper[4751]: I0316 00:19:44.677941 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"40097d75-fb3f-4891-8f6f-f90d72ee8ca0","Type":"ContainerDied","Data":"1a4c6d5f84b581922957d9ffe129a1b197b6a75509a592c116d7a9dfb1599faf"} Mar 16 00:19:44 crc kubenswrapper[4751]: I0316 00:19:44.720019 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-1-build_40097d75-fb3f-4891-8f6f-f90d72ee8ca0/manage-dockerfile/0.log" Mar 16 00:19:45 crc kubenswrapper[4751]: I0316 00:19:45.692596 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"40097d75-fb3f-4891-8f6f-f90d72ee8ca0","Type":"ContainerStarted","Data":"8bbce582b4707b1e3c54998cf6a0c24ee898ee119cadf5d5c44504554573591e"} Mar 16 00:19:45 crc kubenswrapper[4751]: I0316 00:19:45.732428 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-framework-index-1-build" podStartSLOduration=5.288512455 podStartE2EDuration="11.732402192s" podCreationTimestamp="2026-03-16 00:19:34 +0000 UTC" firstStartedPulling="2026-03-16 00:19:35.087618121 +0000 UTC m=+803.356950421" lastFinishedPulling="2026-03-16 00:19:41.531507878 +0000 UTC m=+809.800840158" observedRunningTime="2026-03-16 00:19:45.72877681 +0000 UTC m=+813.998109160" watchObservedRunningTime="2026-03-16 00:19:45.732402192 +0000 UTC m=+814.001734522" Mar 16 00:19:49 crc kubenswrapper[4751]: I0316 00:19:49.071144 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="f222ef67-630e-4203-8857-b2d4fb4f1e89" containerName="elasticsearch" probeResult="failure" output=< Mar 16 00:19:49 crc kubenswrapper[4751]: {"timestamp": "2026-03-16T00:19:49+00:00", "message": "readiness probe failed", "curl_rc": "7"} Mar 16 00:19:49 crc kubenswrapper[4751]: > Mar 16 00:19:54 crc kubenswrapper[4751]: I0316 00:19:54.326458 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.139166 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560340-bbd76"] Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.140681 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560340-bbd76" Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.142932 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.143195 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.143633 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.154421 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560340-bbd76"] Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.156147 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59tbs\" (UniqueName: \"kubernetes.io/projected/3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc-kube-api-access-59tbs\") pod \"auto-csr-approver-29560340-bbd76\" (UID: \"3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc\") " pod="openshift-infra/auto-csr-approver-29560340-bbd76" Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.257377 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59tbs\" (UniqueName: \"kubernetes.io/projected/3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc-kube-api-access-59tbs\") pod \"auto-csr-approver-29560340-bbd76\" (UID: \"3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc\") " pod="openshift-infra/auto-csr-approver-29560340-bbd76" Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.290464 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59tbs\" (UniqueName: \"kubernetes.io/projected/3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc-kube-api-access-59tbs\") pod \"auto-csr-approver-29560340-bbd76\" (UID: \"3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc\") " pod="openshift-infra/auto-csr-approver-29560340-bbd76" Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.468004 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560340-bbd76" Mar 16 00:20:00 crc kubenswrapper[4751]: I0316 00:20:00.812579 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560340-bbd76"] Mar 16 00:20:00 crc kubenswrapper[4751]: W0316 00:20:00.824228 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d7f31dc_1e3a_456c_b00c_d1f8d87e99fc.slice/crio-cf60964c443a3e4d32af43b61533b1896c3f3b13707ad31a47c66b453e955bc8 WatchSource:0}: Error finding container cf60964c443a3e4d32af43b61533b1896c3f3b13707ad31a47c66b453e955bc8: Status 404 returned error can't find the container with id cf60964c443a3e4d32af43b61533b1896c3f3b13707ad31a47c66b453e955bc8 Mar 16 00:20:01 crc kubenswrapper[4751]: I0316 00:20:01.800688 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560340-bbd76" event={"ID":"3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc","Type":"ContainerStarted","Data":"cf60964c443a3e4d32af43b61533b1896c3f3b13707ad31a47c66b453e955bc8"} Mar 16 00:20:03 crc kubenswrapper[4751]: I0316 00:20:03.816344 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560340-bbd76" event={"ID":"3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc","Type":"ContainerStarted","Data":"ef49699cabe449908208ec67e1ac110ada3edeb36c1b231b41f28a4c7af80c62"} Mar 16 00:20:03 crc kubenswrapper[4751]: I0316 00:20:03.832174 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29560340-bbd76" podStartSLOduration=1.173769904 podStartE2EDuration="3.83215392s" podCreationTimestamp="2026-03-16 00:20:00 +0000 UTC" firstStartedPulling="2026-03-16 00:20:00.835251306 +0000 UTC m=+829.104583606" lastFinishedPulling="2026-03-16 00:20:03.493635332 +0000 UTC m=+831.762967622" observedRunningTime="2026-03-16 00:20:03.827452338 +0000 UTC m=+832.096784638" watchObservedRunningTime="2026-03-16 00:20:03.83215392 +0000 UTC m=+832.101486220" Mar 16 00:20:04 crc kubenswrapper[4751]: I0316 00:20:04.823261 4751 generic.go:334] "Generic (PLEG): container finished" podID="3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc" containerID="ef49699cabe449908208ec67e1ac110ada3edeb36c1b231b41f28a4c7af80c62" exitCode=0 Mar 16 00:20:04 crc kubenswrapper[4751]: I0316 00:20:04.823319 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560340-bbd76" event={"ID":"3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc","Type":"ContainerDied","Data":"ef49699cabe449908208ec67e1ac110ada3edeb36c1b231b41f28a4c7af80c62"} Mar 16 00:20:06 crc kubenswrapper[4751]: I0316 00:20:06.116225 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560340-bbd76" Mar 16 00:20:06 crc kubenswrapper[4751]: I0316 00:20:06.246305 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59tbs\" (UniqueName: \"kubernetes.io/projected/3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc-kube-api-access-59tbs\") pod \"3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc\" (UID: \"3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc\") " Mar 16 00:20:06 crc kubenswrapper[4751]: I0316 00:20:06.255425 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc-kube-api-access-59tbs" (OuterVolumeSpecName: "kube-api-access-59tbs") pod "3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc" (UID: "3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc"). InnerVolumeSpecName "kube-api-access-59tbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:20:06 crc kubenswrapper[4751]: I0316 00:20:06.347224 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59tbs\" (UniqueName: \"kubernetes.io/projected/3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc-kube-api-access-59tbs\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:06 crc kubenswrapper[4751]: I0316 00:20:06.839991 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560340-bbd76" event={"ID":"3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc","Type":"ContainerDied","Data":"cf60964c443a3e4d32af43b61533b1896c3f3b13707ad31a47c66b453e955bc8"} Mar 16 00:20:06 crc kubenswrapper[4751]: I0316 00:20:06.840028 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf60964c443a3e4d32af43b61533b1896c3f3b13707ad31a47c66b453e955bc8" Mar 16 00:20:06 crc kubenswrapper[4751]: I0316 00:20:06.840048 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560340-bbd76" Mar 16 00:20:06 crc kubenswrapper[4751]: I0316 00:20:06.909391 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29560334-5w59w"] Mar 16 00:20:06 crc kubenswrapper[4751]: I0316 00:20:06.918055 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29560334-5w59w"] Mar 16 00:20:08 crc kubenswrapper[4751]: I0316 00:20:08.465272 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa4a917-7cee-46f0-ad11-1fcdeacb0469" path="/var/lib/kubelet/pods/6aa4a917-7cee-46f0-ad11-1fcdeacb0469/volumes" Mar 16 00:20:13 crc kubenswrapper[4751]: I0316 00:20:13.172040 4751 scope.go:117] "RemoveContainer" containerID="2bff3cf5eda1ea9b616976a38f61facd57dfd144f2a9db25cc1700d1a1d9bd03" Mar 16 00:20:26 crc kubenswrapper[4751]: I0316 00:20:26.988638 4751 generic.go:334] "Generic (PLEG): container finished" podID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerID="8bbce582b4707b1e3c54998cf6a0c24ee898ee119cadf5d5c44504554573591e" exitCode=0 Mar 16 00:20:26 crc kubenswrapper[4751]: I0316 00:20:26.989387 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"40097d75-fb3f-4891-8f6f-f90d72ee8ca0","Type":"ContainerDied","Data":"8bbce582b4707b1e3c54998cf6a0c24ee898ee119cadf5d5c44504554573591e"} Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.366666 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.419230 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-node-pullsecrets\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.419426 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-ca-bundles\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.420365 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.421413 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-root\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.421551 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-2bxtc-push\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-push\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.421713 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-proxy-ca-bundles\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.421789 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-blob-cache\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.421847 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.421864 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-system-configs\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.421994 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwq7w\" (UniqueName: \"kubernetes.io/projected/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-kube-api-access-fwq7w\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.422038 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildworkdir\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.422172 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.422238 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-run\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.422289 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildcachedir\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.422842 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-2bxtc-pull\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-pull\") pod \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\" (UID: \"40097d75-fb3f-4891-8f6f-f90d72ee8ca0\") " Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.423176 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.423880 4751 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.423922 4751 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.423953 4751 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.427003 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.427703 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.428242 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.430611 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.431741 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-push" (OuterVolumeSpecName: "builder-dockercfg-2bxtc-push") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "builder-dockercfg-2bxtc-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.431993 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-kube-api-access-fwq7w" (OuterVolumeSpecName: "kube-api-access-fwq7w") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "kube-api-access-fwq7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.433012 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-service-telemetry-framework-index-dockercfg-user-build-volume" (OuterVolumeSpecName: "service-telemetry-framework-index-dockercfg-user-build-volume") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "service-telemetry-framework-index-dockercfg-user-build-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.450081 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-pull" (OuterVolumeSpecName: "builder-dockercfg-2bxtc-pull") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "builder-dockercfg-2bxtc-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.525646 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwq7w\" (UniqueName: \"kubernetes.io/projected/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-kube-api-access-fwq7w\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.525683 4751 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.525696 4751 reconciler_common.go:293] "Volume detached for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-service-telemetry-framework-index-dockercfg-user-build-volume\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.525710 4751 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.525726 4751 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.525737 4751 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-2bxtc-pull\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-pull\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.525750 4751 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-2bxtc-push\" (UniqueName: \"kubernetes.io/secret/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-builder-dockercfg-2bxtc-push\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.525763 4751 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.633330 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:28 crc kubenswrapper[4751]: I0316 00:20:28.727634 4751 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.008343 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"40097d75-fb3f-4891-8f6f-f90d72ee8ca0","Type":"ContainerDied","Data":"0047df3baab30eea47b72c3c73032018e0943a56a133f3e81d50a38c09fee600"} Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.008382 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0047df3baab30eea47b72c3c73032018e0943a56a133f3e81d50a38c09fee600" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.008488 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.735896 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "40097d75-fb3f-4891-8f6f-f90d72ee8ca0" (UID: "40097d75-fb3f-4891-8f6f-f90d72ee8ca0"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.745071 4751 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/40097d75-fb3f-4891-8f6f-f90d72ee8ca0-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.932985 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-t9gcf"] Mar 16 00:20:29 crc kubenswrapper[4751]: E0316 00:20:29.933376 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc" containerName="oc" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.933405 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc" containerName="oc" Mar 16 00:20:29 crc kubenswrapper[4751]: E0316 00:20:29.933435 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerName="manage-dockerfile" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.933451 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerName="manage-dockerfile" Mar 16 00:20:29 crc kubenswrapper[4751]: E0316 00:20:29.933493 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerName="git-clone" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.933513 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerName="git-clone" Mar 16 00:20:29 crc kubenswrapper[4751]: E0316 00:20:29.933547 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerName="docker-build" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.933563 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerName="docker-build" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.933797 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc" containerName="oc" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.933829 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="40097d75-fb3f-4891-8f6f-f90d72ee8ca0" containerName="docker-build" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.934651 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-t9gcf" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.939236 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"infrawatch-operators-dockercfg-mcsdv" Mar 16 00:20:29 crc kubenswrapper[4751]: I0316 00:20:29.946146 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-t9gcf"] Mar 16 00:20:30 crc kubenswrapper[4751]: I0316 00:20:30.049058 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqhw7\" (UniqueName: \"kubernetes.io/projected/195d2190-b3fc-46ec-81af-0f0ceda18d9b-kube-api-access-hqhw7\") pod \"infrawatch-operators-t9gcf\" (UID: \"195d2190-b3fc-46ec-81af-0f0ceda18d9b\") " pod="service-telemetry/infrawatch-operators-t9gcf" Mar 16 00:20:30 crc kubenswrapper[4751]: I0316 00:20:30.151476 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqhw7\" (UniqueName: \"kubernetes.io/projected/195d2190-b3fc-46ec-81af-0f0ceda18d9b-kube-api-access-hqhw7\") pod \"infrawatch-operators-t9gcf\" (UID: \"195d2190-b3fc-46ec-81af-0f0ceda18d9b\") " pod="service-telemetry/infrawatch-operators-t9gcf" Mar 16 00:20:30 crc kubenswrapper[4751]: I0316 00:20:30.201665 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqhw7\" (UniqueName: \"kubernetes.io/projected/195d2190-b3fc-46ec-81af-0f0ceda18d9b-kube-api-access-hqhw7\") pod \"infrawatch-operators-t9gcf\" (UID: \"195d2190-b3fc-46ec-81af-0f0ceda18d9b\") " pod="service-telemetry/infrawatch-operators-t9gcf" Mar 16 00:20:30 crc kubenswrapper[4751]: I0316 00:20:30.263232 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-t9gcf" Mar 16 00:20:30 crc kubenswrapper[4751]: I0316 00:20:30.526035 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-t9gcf"] Mar 16 00:20:30 crc kubenswrapper[4751]: W0316 00:20:30.537450 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod195d2190_b3fc_46ec_81af_0f0ceda18d9b.slice/crio-3e9e348eec3abd2ab93822e728a31e8bdb21f1bf60402724a766da1d4c58ec1c WatchSource:0}: Error finding container 3e9e348eec3abd2ab93822e728a31e8bdb21f1bf60402724a766da1d4c58ec1c: Status 404 returned error can't find the container with id 3e9e348eec3abd2ab93822e728a31e8bdb21f1bf60402724a766da1d4c58ec1c Mar 16 00:20:31 crc kubenswrapper[4751]: I0316 00:20:31.023158 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-t9gcf" event={"ID":"195d2190-b3fc-46ec-81af-0f0ceda18d9b","Type":"ContainerStarted","Data":"3e9e348eec3abd2ab93822e728a31e8bdb21f1bf60402724a766da1d4c58ec1c"} Mar 16 00:20:34 crc kubenswrapper[4751]: I0316 00:20:34.301481 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-t9gcf"] Mar 16 00:20:35 crc kubenswrapper[4751]: I0316 00:20:35.123821 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-xx5lr"] Mar 16 00:20:35 crc kubenswrapper[4751]: I0316 00:20:35.125347 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-xx5lr" Mar 16 00:20:35 crc kubenswrapper[4751]: I0316 00:20:35.134414 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-xx5lr"] Mar 16 00:20:35 crc kubenswrapper[4751]: I0316 00:20:35.230317 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4crn\" (UniqueName: \"kubernetes.io/projected/66aad594-4594-4f54-9010-2ba72be6ec67-kube-api-access-m4crn\") pod \"infrawatch-operators-xx5lr\" (UID: \"66aad594-4594-4f54-9010-2ba72be6ec67\") " pod="service-telemetry/infrawatch-operators-xx5lr" Mar 16 00:20:35 crc kubenswrapper[4751]: I0316 00:20:35.332039 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4crn\" (UniqueName: \"kubernetes.io/projected/66aad594-4594-4f54-9010-2ba72be6ec67-kube-api-access-m4crn\") pod \"infrawatch-operators-xx5lr\" (UID: \"66aad594-4594-4f54-9010-2ba72be6ec67\") " pod="service-telemetry/infrawatch-operators-xx5lr" Mar 16 00:20:35 crc kubenswrapper[4751]: I0316 00:20:35.370932 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4crn\" (UniqueName: \"kubernetes.io/projected/66aad594-4594-4f54-9010-2ba72be6ec67-kube-api-access-m4crn\") pod \"infrawatch-operators-xx5lr\" (UID: \"66aad594-4594-4f54-9010-2ba72be6ec67\") " pod="service-telemetry/infrawatch-operators-xx5lr" Mar 16 00:20:35 crc kubenswrapper[4751]: I0316 00:20:35.460895 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-xx5lr" Mar 16 00:20:41 crc kubenswrapper[4751]: I0316 00:20:41.657332 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-xx5lr"] Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.109013 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-xx5lr" event={"ID":"66aad594-4594-4f54-9010-2ba72be6ec67","Type":"ContainerStarted","Data":"3fd225349b77a6cd7bf1b016f337c5f0aa70d9406dee8bc33ce236aed3311dc8"} Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.109073 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-xx5lr" event={"ID":"66aad594-4594-4f54-9010-2ba72be6ec67","Type":"ContainerStarted","Data":"07f08b65491517aac25fe96bdb5663cd20028e8005082ecea4dd026ccc3c2381"} Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.112594 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-t9gcf" event={"ID":"195d2190-b3fc-46ec-81af-0f0ceda18d9b","Type":"ContainerStarted","Data":"9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551"} Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.112773 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-t9gcf" podUID="195d2190-b3fc-46ec-81af-0f0ceda18d9b" containerName="registry-server" containerID="cri-o://9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551" gracePeriod=2 Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.130664 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-xx5lr" podStartSLOduration=7.001968863 podStartE2EDuration="7.130637045s" podCreationTimestamp="2026-03-16 00:20:35 +0000 UTC" firstStartedPulling="2026-03-16 00:20:41.671346157 +0000 UTC m=+869.940678477" lastFinishedPulling="2026-03-16 00:20:41.800014359 +0000 UTC m=+870.069346659" observedRunningTime="2026-03-16 00:20:42.128129225 +0000 UTC m=+870.397461535" watchObservedRunningTime="2026-03-16 00:20:42.130637045 +0000 UTC m=+870.399969375" Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.158479 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-t9gcf" podStartSLOduration=2.357666076 podStartE2EDuration="13.158456478s" podCreationTimestamp="2026-03-16 00:20:29 +0000 UTC" firstStartedPulling="2026-03-16 00:20:30.539675773 +0000 UTC m=+858.809008073" lastFinishedPulling="2026-03-16 00:20:41.340466135 +0000 UTC m=+869.609798475" observedRunningTime="2026-03-16 00:20:42.151986016 +0000 UTC m=+870.421318336" watchObservedRunningTime="2026-03-16 00:20:42.158456478 +0000 UTC m=+870.427788788" Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.559045 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-t9gcf" Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.631358 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqhw7\" (UniqueName: \"kubernetes.io/projected/195d2190-b3fc-46ec-81af-0f0ceda18d9b-kube-api-access-hqhw7\") pod \"195d2190-b3fc-46ec-81af-0f0ceda18d9b\" (UID: \"195d2190-b3fc-46ec-81af-0f0ceda18d9b\") " Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.640100 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/195d2190-b3fc-46ec-81af-0f0ceda18d9b-kube-api-access-hqhw7" (OuterVolumeSpecName: "kube-api-access-hqhw7") pod "195d2190-b3fc-46ec-81af-0f0ceda18d9b" (UID: "195d2190-b3fc-46ec-81af-0f0ceda18d9b"). InnerVolumeSpecName "kube-api-access-hqhw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:20:42 crc kubenswrapper[4751]: I0316 00:20:42.733895 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqhw7\" (UniqueName: \"kubernetes.io/projected/195d2190-b3fc-46ec-81af-0f0ceda18d9b-kube-api-access-hqhw7\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:43 crc kubenswrapper[4751]: I0316 00:20:43.123566 4751 generic.go:334] "Generic (PLEG): container finished" podID="195d2190-b3fc-46ec-81af-0f0ceda18d9b" containerID="9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551" exitCode=0 Mar 16 00:20:43 crc kubenswrapper[4751]: I0316 00:20:43.123646 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-t9gcf" event={"ID":"195d2190-b3fc-46ec-81af-0f0ceda18d9b","Type":"ContainerDied","Data":"9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551"} Mar 16 00:20:43 crc kubenswrapper[4751]: I0316 00:20:43.123696 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-t9gcf" Mar 16 00:20:43 crc kubenswrapper[4751]: I0316 00:20:43.123818 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-t9gcf" event={"ID":"195d2190-b3fc-46ec-81af-0f0ceda18d9b","Type":"ContainerDied","Data":"3e9e348eec3abd2ab93822e728a31e8bdb21f1bf60402724a766da1d4c58ec1c"} Mar 16 00:20:43 crc kubenswrapper[4751]: I0316 00:20:43.125235 4751 scope.go:117] "RemoveContainer" containerID="9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551" Mar 16 00:20:43 crc kubenswrapper[4751]: I0316 00:20:43.146575 4751 scope.go:117] "RemoveContainer" containerID="9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551" Mar 16 00:20:43 crc kubenswrapper[4751]: E0316 00:20:43.147034 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551\": container with ID starting with 9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551 not found: ID does not exist" containerID="9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551" Mar 16 00:20:43 crc kubenswrapper[4751]: I0316 00:20:43.147072 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551"} err="failed to get container status \"9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551\": rpc error: code = NotFound desc = could not find container \"9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551\": container with ID starting with 9177ec1776bdc5a6941034537570f3c6437c68b7be7b759bc69d8ea732a5d551 not found: ID does not exist" Mar 16 00:20:43 crc kubenswrapper[4751]: I0316 00:20:43.180636 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-t9gcf"] Mar 16 00:20:43 crc kubenswrapper[4751]: I0316 00:20:43.188890 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-t9gcf"] Mar 16 00:20:44 crc kubenswrapper[4751]: I0316 00:20:44.464711 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="195d2190-b3fc-46ec-81af-0f0ceda18d9b" path="/var/lib/kubelet/pods/195d2190-b3fc-46ec-81af-0f0ceda18d9b/volumes" Mar 16 00:20:45 crc kubenswrapper[4751]: I0316 00:20:45.461742 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-xx5lr" Mar 16 00:20:45 crc kubenswrapper[4751]: I0316 00:20:45.461860 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-xx5lr" Mar 16 00:20:45 crc kubenswrapper[4751]: I0316 00:20:45.504573 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-xx5lr" Mar 16 00:20:46 crc kubenswrapper[4751]: I0316 00:20:46.186713 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-xx5lr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.578502 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr"] Mar 16 00:20:48 crc kubenswrapper[4751]: E0316 00:20:48.579267 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="195d2190-b3fc-46ec-81af-0f0ceda18d9b" containerName="registry-server" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.579298 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="195d2190-b3fc-46ec-81af-0f0ceda18d9b" containerName="registry-server" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.579500 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="195d2190-b3fc-46ec-81af-0f0ceda18d9b" containerName="registry-server" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.580877 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.592567 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr"] Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.619598 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-util\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.619673 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh575\" (UniqueName: \"kubernetes.io/projected/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-kube-api-access-gh575\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.619747 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-bundle\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.722254 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-bundle\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.722411 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-util\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.722562 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh575\" (UniqueName: \"kubernetes.io/projected/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-kube-api-access-gh575\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.723293 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-util\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.723301 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-bundle\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.761791 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh575\" (UniqueName: \"kubernetes.io/projected/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-kube-api-access-gh575\") pod \"27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:48 crc kubenswrapper[4751]: I0316 00:20:48.908525 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.188272 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr"] Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.347347 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8"] Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.348754 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.356575 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8"] Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.432420 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-util\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.432524 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-bundle\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.432579 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sg6k\" (UniqueName: \"kubernetes.io/projected/7c775998-972f-4ae3-a0c1-0f1036008c5f-kube-api-access-9sg6k\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.533963 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sg6k\" (UniqueName: \"kubernetes.io/projected/7c775998-972f-4ae3-a0c1-0f1036008c5f-kube-api-access-9sg6k\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.534224 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-util\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.534331 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-bundle\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.535245 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-bundle\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.535371 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-util\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.571152 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sg6k\" (UniqueName: \"kubernetes.io/projected/7c775998-972f-4ae3-a0c1-0f1036008c5f-kube-api-access-9sg6k\") pod \"cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.666990 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:49 crc kubenswrapper[4751]: I0316 00:20:49.949039 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8"] Mar 16 00:20:49 crc kubenswrapper[4751]: W0316 00:20:49.949526 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c775998_972f_4ae3_a0c1_0f1036008c5f.slice/crio-d1a25628913b68ddf464db2ed6e38cfdfbb3a2bf3d3fdc778dacc5036c88240b WatchSource:0}: Error finding container d1a25628913b68ddf464db2ed6e38cfdfbb3a2bf3d3fdc778dacc5036c88240b: Status 404 returned error can't find the container with id d1a25628913b68ddf464db2ed6e38cfdfbb3a2bf3d3fdc778dacc5036c88240b Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.166167 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz"] Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.167543 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.171567 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.181592 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" event={"ID":"7c775998-972f-4ae3-a0c1-0f1036008c5f","Type":"ContainerStarted","Data":"d1a25628913b68ddf464db2ed6e38cfdfbb3a2bf3d3fdc778dacc5036c88240b"} Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.181659 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz"] Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.183124 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" event={"ID":"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3","Type":"ContainerStarted","Data":"387c025ae60147638e6b2ea51374e14b037c3fc4f56b2302a6b362e5fb2b11cc"} Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.246817 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.246916 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pjpv\" (UniqueName: \"kubernetes.io/projected/6c930011-0311-4778-a108-885ea6ecd7ec-kube-api-access-5pjpv\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.246976 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.348469 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pjpv\" (UniqueName: \"kubernetes.io/projected/6c930011-0311-4778-a108-885ea6ecd7ec-kube-api-access-5pjpv\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.348563 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.348605 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.349372 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.350032 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.376982 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pjpv\" (UniqueName: \"kubernetes.io/projected/6c930011-0311-4778-a108-885ea6ecd7ec-kube-api-access-5pjpv\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.539532 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:50 crc kubenswrapper[4751]: I0316 00:20:50.747233 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz"] Mar 16 00:20:50 crc kubenswrapper[4751]: W0316 00:20:50.755599 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c930011_0311_4778_a108_885ea6ecd7ec.slice/crio-fa58fac301d11f36c54a74b8668080c3e0cb391a058bb7d899eef8330471f9d0 WatchSource:0}: Error finding container fa58fac301d11f36c54a74b8668080c3e0cb391a058bb7d899eef8330471f9d0: Status 404 returned error can't find the container with id fa58fac301d11f36c54a74b8668080c3e0cb391a058bb7d899eef8330471f9d0 Mar 16 00:20:51 crc kubenswrapper[4751]: I0316 00:20:51.191763 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" event={"ID":"6c930011-0311-4778-a108-885ea6ecd7ec","Type":"ContainerStarted","Data":"fa58fac301d11f36c54a74b8668080c3e0cb391a058bb7d899eef8330471f9d0"} Mar 16 00:20:52 crc kubenswrapper[4751]: I0316 00:20:52.203887 4751 generic.go:334] "Generic (PLEG): container finished" podID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerID="cefbd68bc4e694f813676d72bfbe5911c21c8242c93e2476a3a0b3186ac70d50" exitCode=0 Mar 16 00:20:52 crc kubenswrapper[4751]: I0316 00:20:52.203972 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" event={"ID":"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3","Type":"ContainerDied","Data":"cefbd68bc4e694f813676d72bfbe5911c21c8242c93e2476a3a0b3186ac70d50"} Mar 16 00:20:52 crc kubenswrapper[4751]: I0316 00:20:52.206547 4751 generic.go:334] "Generic (PLEG): container finished" podID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerID="84496d55c9f62b0cc4648ae8211611ce34402f4b47a198a9e255c9ab552d37ac" exitCode=0 Mar 16 00:20:52 crc kubenswrapper[4751]: I0316 00:20:52.206627 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" event={"ID":"7c775998-972f-4ae3-a0c1-0f1036008c5f","Type":"ContainerDied","Data":"84496d55c9f62b0cc4648ae8211611ce34402f4b47a198a9e255c9ab552d37ac"} Mar 16 00:20:52 crc kubenswrapper[4751]: I0316 00:20:52.213133 4751 generic.go:334] "Generic (PLEG): container finished" podID="6c930011-0311-4778-a108-885ea6ecd7ec" containerID="f6d2bb4d263273617b3ee79bf10fd6d9cddaa9563673f9216ce85d2f45cb06b1" exitCode=0 Mar 16 00:20:52 crc kubenswrapper[4751]: I0316 00:20:52.213194 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" event={"ID":"6c930011-0311-4778-a108-885ea6ecd7ec","Type":"ContainerDied","Data":"f6d2bb4d263273617b3ee79bf10fd6d9cddaa9563673f9216ce85d2f45cb06b1"} Mar 16 00:20:53 crc kubenswrapper[4751]: I0316 00:20:53.238956 4751 generic.go:334] "Generic (PLEG): container finished" podID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerID="cfab428ad4527fce1ed65689961ebe76ff3a5ce464ac35b3f24e5727f31ab93d" exitCode=0 Mar 16 00:20:53 crc kubenswrapper[4751]: I0316 00:20:53.239029 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" event={"ID":"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3","Type":"ContainerDied","Data":"cfab428ad4527fce1ed65689961ebe76ff3a5ce464ac35b3f24e5727f31ab93d"} Mar 16 00:20:53 crc kubenswrapper[4751]: I0316 00:20:53.245733 4751 generic.go:334] "Generic (PLEG): container finished" podID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerID="88ee4294eb2b6e2596ff93a092aca1cbc27232002bcebfb9ab88d4ad021ee255" exitCode=0 Mar 16 00:20:53 crc kubenswrapper[4751]: I0316 00:20:53.245804 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" event={"ID":"7c775998-972f-4ae3-a0c1-0f1036008c5f","Type":"ContainerDied","Data":"88ee4294eb2b6e2596ff93a092aca1cbc27232002bcebfb9ab88d4ad021ee255"} Mar 16 00:20:54 crc kubenswrapper[4751]: I0316 00:20:54.256050 4751 generic.go:334] "Generic (PLEG): container finished" podID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerID="621d8807e5e33a4e422eea4001cfefc2c479ce07e05813c242499120382bfe17" exitCode=0 Mar 16 00:20:54 crc kubenswrapper[4751]: I0316 00:20:54.256199 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" event={"ID":"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3","Type":"ContainerDied","Data":"621d8807e5e33a4e422eea4001cfefc2c479ce07e05813c242499120382bfe17"} Mar 16 00:20:54 crc kubenswrapper[4751]: I0316 00:20:54.259696 4751 generic.go:334] "Generic (PLEG): container finished" podID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerID="f0bfef96d1dadb78561ee254a24dfcfc460d7df8a68703ef2984ab0d5bd7b389" exitCode=0 Mar 16 00:20:54 crc kubenswrapper[4751]: I0316 00:20:54.259815 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" event={"ID":"7c775998-972f-4ae3-a0c1-0f1036008c5f","Type":"ContainerDied","Data":"f0bfef96d1dadb78561ee254a24dfcfc460d7df8a68703ef2984ab0d5bd7b389"} Mar 16 00:20:54 crc kubenswrapper[4751]: I0316 00:20:54.262619 4751 generic.go:334] "Generic (PLEG): container finished" podID="6c930011-0311-4778-a108-885ea6ecd7ec" containerID="1bf9bc77b94449f5feca8112b5739e3de06269764940f550ae0b2fd3b72a502e" exitCode=0 Mar 16 00:20:54 crc kubenswrapper[4751]: I0316 00:20:54.262660 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" event={"ID":"6c930011-0311-4778-a108-885ea6ecd7ec","Type":"ContainerDied","Data":"1bf9bc77b94449f5feca8112b5739e3de06269764940f550ae0b2fd3b72a502e"} Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.274308 4751 generic.go:334] "Generic (PLEG): container finished" podID="6c930011-0311-4778-a108-885ea6ecd7ec" containerID="85e49b631075426cad82736e568cfb8960b163650c18a87902e645c24bde7938" exitCode=0 Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.274383 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" event={"ID":"6c930011-0311-4778-a108-885ea6ecd7ec","Type":"ContainerDied","Data":"85e49b631075426cad82736e568cfb8960b163650c18a87902e645c24bde7938"} Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.607561 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.611554 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.729767 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-util\") pod \"7c775998-972f-4ae3-a0c1-0f1036008c5f\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.729872 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-bundle\") pod \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.729942 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-util\") pod \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.730023 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh575\" (UniqueName: \"kubernetes.io/projected/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-kube-api-access-gh575\") pod \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\" (UID: \"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3\") " Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.730068 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-bundle\") pod \"7c775998-972f-4ae3-a0c1-0f1036008c5f\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.730156 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sg6k\" (UniqueName: \"kubernetes.io/projected/7c775998-972f-4ae3-a0c1-0f1036008c5f-kube-api-access-9sg6k\") pod \"7c775998-972f-4ae3-a0c1-0f1036008c5f\" (UID: \"7c775998-972f-4ae3-a0c1-0f1036008c5f\") " Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.730527 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-bundle" (OuterVolumeSpecName: "bundle") pod "478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" (UID: "478dd9e5-4178-46d6-a0c1-0e6603a6f6e3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.730733 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-bundle" (OuterVolumeSpecName: "bundle") pod "7c775998-972f-4ae3-a0c1-0f1036008c5f" (UID: "7c775998-972f-4ae3-a0c1-0f1036008c5f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.735170 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-kube-api-access-gh575" (OuterVolumeSpecName: "kube-api-access-gh575") pod "478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" (UID: "478dd9e5-4178-46d6-a0c1-0e6603a6f6e3"). InnerVolumeSpecName "kube-api-access-gh575". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.735342 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c775998-972f-4ae3-a0c1-0f1036008c5f-kube-api-access-9sg6k" (OuterVolumeSpecName: "kube-api-access-9sg6k") pod "7c775998-972f-4ae3-a0c1-0f1036008c5f" (UID: "7c775998-972f-4ae3-a0c1-0f1036008c5f"). InnerVolumeSpecName "kube-api-access-9sg6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.748130 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-util" (OuterVolumeSpecName: "util") pod "478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" (UID: "478dd9e5-4178-46d6-a0c1-0e6603a6f6e3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.755157 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-util" (OuterVolumeSpecName: "util") pod "7c775998-972f-4ae3-a0c1-0f1036008c5f" (UID: "7c775998-972f-4ae3-a0c1-0f1036008c5f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.831476 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sg6k\" (UniqueName: \"kubernetes.io/projected/7c775998-972f-4ae3-a0c1-0f1036008c5f-kube-api-access-9sg6k\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.831513 4751 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-util\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.831528 4751 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.831538 4751 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-util\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.831549 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh575\" (UniqueName: \"kubernetes.io/projected/478dd9e5-4178-46d6-a0c1-0e6603a6f6e3-kube-api-access-gh575\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:55 crc kubenswrapper[4751]: I0316 00:20:55.831560 4751 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c775998-972f-4ae3-a0c1-0f1036008c5f-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.289719 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" event={"ID":"478dd9e5-4178-46d6-a0c1-0e6603a6f6e3","Type":"ContainerDied","Data":"387c025ae60147638e6b2ea51374e14b037c3fc4f56b2302a6b362e5fb2b11cc"} Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.289768 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/27e4a3b82b847aaaac340f98fd9ec51c99f28242b589c6c251a26fbc3bs8qrr" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.289790 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="387c025ae60147638e6b2ea51374e14b037c3fc4f56b2302a6b362e5fb2b11cc" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.299633 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" event={"ID":"7c775998-972f-4ae3-a0c1-0f1036008c5f","Type":"ContainerDied","Data":"d1a25628913b68ddf464db2ed6e38cfdfbb3a2bf3d3fdc778dacc5036c88240b"} Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.299684 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1a25628913b68ddf464db2ed6e38cfdfbb3a2bf3d3fdc778dacc5036c88240b" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.299726 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/cd07ebce3b618aaffa8e106dab3e8eb93287fbb0e4c5a9c0f6ea8fc7eb5rsc8" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.623805 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.775281 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-bundle\") pod \"6c930011-0311-4778-a108-885ea6ecd7ec\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.775397 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pjpv\" (UniqueName: \"kubernetes.io/projected/6c930011-0311-4778-a108-885ea6ecd7ec-kube-api-access-5pjpv\") pod \"6c930011-0311-4778-a108-885ea6ecd7ec\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.775437 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-util\") pod \"6c930011-0311-4778-a108-885ea6ecd7ec\" (UID: \"6c930011-0311-4778-a108-885ea6ecd7ec\") " Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.775929 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-bundle" (OuterVolumeSpecName: "bundle") pod "6c930011-0311-4778-a108-885ea6ecd7ec" (UID: "6c930011-0311-4778-a108-885ea6ecd7ec"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.781093 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c930011-0311-4778-a108-885ea6ecd7ec-kube-api-access-5pjpv" (OuterVolumeSpecName: "kube-api-access-5pjpv") pod "6c930011-0311-4778-a108-885ea6ecd7ec" (UID: "6c930011-0311-4778-a108-885ea6ecd7ec"). InnerVolumeSpecName "kube-api-access-5pjpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.861128 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-util" (OuterVolumeSpecName: "util") pod "6c930011-0311-4778-a108-885ea6ecd7ec" (UID: "6c930011-0311-4778-a108-885ea6ecd7ec"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.876976 4751 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-bundle\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.877019 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pjpv\" (UniqueName: \"kubernetes.io/projected/6c930011-0311-4778-a108-885ea6ecd7ec-kube-api-access-5pjpv\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:56 crc kubenswrapper[4751]: I0316 00:20:56.877038 4751 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c930011-0311-4778-a108-885ea6ecd7ec-util\") on node \"crc\" DevicePath \"\"" Mar 16 00:20:57 crc kubenswrapper[4751]: I0316 00:20:57.309449 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" event={"ID":"6c930011-0311-4778-a108-885ea6ecd7ec","Type":"ContainerDied","Data":"fa58fac301d11f36c54a74b8668080c3e0cb391a058bb7d899eef8330471f9d0"} Mar 16 00:20:57 crc kubenswrapper[4751]: I0316 00:20:57.309495 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa58fac301d11f36c54a74b8668080c3e0cb391a058bb7d899eef8330471f9d0" Mar 16 00:20:57 crc kubenswrapper[4751]: I0316 00:20:57.309511 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz" Mar 16 00:20:57 crc kubenswrapper[4751]: I0316 00:20:57.855005 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:20:57 crc kubenswrapper[4751]: I0316 00:20:57.855134 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.355081 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch"] Mar 16 00:21:01 crc kubenswrapper[4751]: E0316 00:21:01.355761 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerName="pull" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.355780 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerName="pull" Mar 16 00:21:01 crc kubenswrapper[4751]: E0316 00:21:01.355796 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerName="extract" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.355806 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerName="extract" Mar 16 00:21:01 crc kubenswrapper[4751]: E0316 00:21:01.355826 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerName="util" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.355837 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerName="util" Mar 16 00:21:01 crc kubenswrapper[4751]: E0316 00:21:01.355849 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerName="pull" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.355860 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerName="pull" Mar 16 00:21:01 crc kubenswrapper[4751]: E0316 00:21:01.356187 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c930011-0311-4778-a108-885ea6ecd7ec" containerName="extract" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.356199 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c930011-0311-4778-a108-885ea6ecd7ec" containerName="extract" Mar 16 00:21:01 crc kubenswrapper[4751]: E0316 00:21:01.356217 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerName="extract" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.356227 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerName="extract" Mar 16 00:21:01 crc kubenswrapper[4751]: E0316 00:21:01.356248 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerName="util" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.356257 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerName="util" Mar 16 00:21:01 crc kubenswrapper[4751]: E0316 00:21:01.356270 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c930011-0311-4778-a108-885ea6ecd7ec" containerName="pull" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.356280 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c930011-0311-4778-a108-885ea6ecd7ec" containerName="pull" Mar 16 00:21:01 crc kubenswrapper[4751]: E0316 00:21:01.356298 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c930011-0311-4778-a108-885ea6ecd7ec" containerName="util" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.356308 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c930011-0311-4778-a108-885ea6ecd7ec" containerName="util" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.356496 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c930011-0311-4778-a108-885ea6ecd7ec" containerName="extract" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.356524 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="478dd9e5-4178-46d6-a0c1-0e6603a6f6e3" containerName="extract" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.356545 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c775998-972f-4ae3-a0c1-0f1036008c5f" containerName="extract" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.357236 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.360007 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wnmw\" (UniqueName: \"kubernetes.io/projected/2d362a30-4cc0-43f3-bae1-482bcc3ad4e5-kube-api-access-8wnmw\") pod \"service-telemetry-operator-fcbbbcfcf-ngbch\" (UID: \"2d362a30-4cc0-43f3-bae1-482bcc3ad4e5\") " pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.360047 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/2d362a30-4cc0-43f3-bae1-482bcc3ad4e5-runner\") pod \"service-telemetry-operator-fcbbbcfcf-ngbch\" (UID: \"2d362a30-4cc0-43f3-bae1-482bcc3ad4e5\") " pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.361620 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-operator-dockercfg-vpmjl" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.368016 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch"] Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.460929 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wnmw\" (UniqueName: \"kubernetes.io/projected/2d362a30-4cc0-43f3-bae1-482bcc3ad4e5-kube-api-access-8wnmw\") pod \"service-telemetry-operator-fcbbbcfcf-ngbch\" (UID: \"2d362a30-4cc0-43f3-bae1-482bcc3ad4e5\") " pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.461007 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/2d362a30-4cc0-43f3-bae1-482bcc3ad4e5-runner\") pod \"service-telemetry-operator-fcbbbcfcf-ngbch\" (UID: \"2d362a30-4cc0-43f3-bae1-482bcc3ad4e5\") " pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.461653 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/2d362a30-4cc0-43f3-bae1-482bcc3ad4e5-runner\") pod \"service-telemetry-operator-fcbbbcfcf-ngbch\" (UID: \"2d362a30-4cc0-43f3-bae1-482bcc3ad4e5\") " pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.497295 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wnmw\" (UniqueName: \"kubernetes.io/projected/2d362a30-4cc0-43f3-bae1-482bcc3ad4e5-kube-api-access-8wnmw\") pod \"service-telemetry-operator-fcbbbcfcf-ngbch\" (UID: \"2d362a30-4cc0-43f3-bae1-482bcc3ad4e5\") " pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" Mar 16 00:21:01 crc kubenswrapper[4751]: I0316 00:21:01.679047 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" Mar 16 00:21:02 crc kubenswrapper[4751]: I0316 00:21:02.160625 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch"] Mar 16 00:21:02 crc kubenswrapper[4751]: W0316 00:21:02.174376 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d362a30_4cc0_43f3_bae1_482bcc3ad4e5.slice/crio-73b245601563982188009bd69cc5bb78de1e65aa937c95df5fb91e0b7d567863 WatchSource:0}: Error finding container 73b245601563982188009bd69cc5bb78de1e65aa937c95df5fb91e0b7d567863: Status 404 returned error can't find the container with id 73b245601563982188009bd69cc5bb78de1e65aa937c95df5fb91e0b7d567863 Mar 16 00:21:02 crc kubenswrapper[4751]: I0316 00:21:02.345632 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" event={"ID":"2d362a30-4cc0-43f3-bae1-482bcc3ad4e5","Type":"ContainerStarted","Data":"73b245601563982188009bd69cc5bb78de1e65aa937c95df5fb91e0b7d567863"} Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.313076 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-697d54f584-rdntd"] Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.313764 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.315398 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-operator-dockercfg-8wkfd" Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.323316 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-697d54f584-rdntd"] Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.487546 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtk9c\" (UniqueName: \"kubernetes.io/projected/6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7-kube-api-access-qtk9c\") pod \"smart-gateway-operator-697d54f584-rdntd\" (UID: \"6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7\") " pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.487631 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7-runner\") pod \"smart-gateway-operator-697d54f584-rdntd\" (UID: \"6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7\") " pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.589544 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7-runner\") pod \"smart-gateway-operator-697d54f584-rdntd\" (UID: \"6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7\") " pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.589819 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtk9c\" (UniqueName: \"kubernetes.io/projected/6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7-kube-api-access-qtk9c\") pod \"smart-gateway-operator-697d54f584-rdntd\" (UID: \"6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7\") " pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.590215 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7-runner\") pod \"smart-gateway-operator-697d54f584-rdntd\" (UID: \"6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7\") " pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.607846 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtk9c\" (UniqueName: \"kubernetes.io/projected/6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7-kube-api-access-qtk9c\") pod \"smart-gateway-operator-697d54f584-rdntd\" (UID: \"6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7\") " pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" Mar 16 00:21:03 crc kubenswrapper[4751]: I0316 00:21:03.635053 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" Mar 16 00:21:04 crc kubenswrapper[4751]: I0316 00:21:04.101535 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-697d54f584-rdntd"] Mar 16 00:21:04 crc kubenswrapper[4751]: W0316 00:21:04.111952 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b438b99_00f6_497f_bfcc_6b2c1d6e0fa7.slice/crio-03d369ef5ebe2d802dec0110a4561430ce9b47d69fd82498396744e3f1f4b221 WatchSource:0}: Error finding container 03d369ef5ebe2d802dec0110a4561430ce9b47d69fd82498396744e3f1f4b221: Status 404 returned error can't find the container with id 03d369ef5ebe2d802dec0110a4561430ce9b47d69fd82498396744e3f1f4b221 Mar 16 00:21:04 crc kubenswrapper[4751]: I0316 00:21:04.379432 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" event={"ID":"6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7","Type":"ContainerStarted","Data":"03d369ef5ebe2d802dec0110a4561430ce9b47d69fd82498396744e3f1f4b221"} Mar 16 00:21:04 crc kubenswrapper[4751]: I0316 00:21:04.738227 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-w5bqg"] Mar 16 00:21:04 crc kubenswrapper[4751]: I0316 00:21:04.738858 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-w5bqg" Mar 16 00:21:04 crc kubenswrapper[4751]: I0316 00:21:04.740745 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-69qzs" Mar 16 00:21:04 crc kubenswrapper[4751]: I0316 00:21:04.763447 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-w5bqg"] Mar 16 00:21:04 crc kubenswrapper[4751]: I0316 00:21:04.811342 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7tjf\" (UniqueName: \"kubernetes.io/projected/cdda1cb0-0c2b-44e1-b1ca-942e8da817aa-kube-api-access-v7tjf\") pod \"interconnect-operator-5bb49f789d-w5bqg\" (UID: \"cdda1cb0-0c2b-44e1-b1ca-942e8da817aa\") " pod="service-telemetry/interconnect-operator-5bb49f789d-w5bqg" Mar 16 00:21:04 crc kubenswrapper[4751]: I0316 00:21:04.912527 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7tjf\" (UniqueName: \"kubernetes.io/projected/cdda1cb0-0c2b-44e1-b1ca-942e8da817aa-kube-api-access-v7tjf\") pod \"interconnect-operator-5bb49f789d-w5bqg\" (UID: \"cdda1cb0-0c2b-44e1-b1ca-942e8da817aa\") " pod="service-telemetry/interconnect-operator-5bb49f789d-w5bqg" Mar 16 00:21:04 crc kubenswrapper[4751]: I0316 00:21:04.934025 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7tjf\" (UniqueName: \"kubernetes.io/projected/cdda1cb0-0c2b-44e1-b1ca-942e8da817aa-kube-api-access-v7tjf\") pod \"interconnect-operator-5bb49f789d-w5bqg\" (UID: \"cdda1cb0-0c2b-44e1-b1ca-942e8da817aa\") " pod="service-telemetry/interconnect-operator-5bb49f789d-w5bqg" Mar 16 00:21:05 crc kubenswrapper[4751]: I0316 00:21:05.056438 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-w5bqg" Mar 16 00:21:05 crc kubenswrapper[4751]: I0316 00:21:05.327710 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-w5bqg"] Mar 16 00:21:05 crc kubenswrapper[4751]: W0316 00:21:05.336413 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdda1cb0_0c2b_44e1_b1ca_942e8da817aa.slice/crio-6c82dee5bcedcf7bef765c388737114d62da16299a5504bb91f776dee7aba5a2 WatchSource:0}: Error finding container 6c82dee5bcedcf7bef765c388737114d62da16299a5504bb91f776dee7aba5a2: Status 404 returned error can't find the container with id 6c82dee5bcedcf7bef765c388737114d62da16299a5504bb91f776dee7aba5a2 Mar 16 00:21:05 crc kubenswrapper[4751]: I0316 00:21:05.392520 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-w5bqg" event={"ID":"cdda1cb0-0c2b-44e1-b1ca-942e8da817aa","Type":"ContainerStarted","Data":"6c82dee5bcedcf7bef765c388737114d62da16299a5504bb91f776dee7aba5a2"} Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.309022 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cl7r6"] Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.310600 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.317806 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cl7r6"] Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.469871 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfrgm\" (UniqueName: \"kubernetes.io/projected/a36702b9-dcfa-4945-80ac-01ac96b0e375-kube-api-access-sfrgm\") pod \"community-operators-cl7r6\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.469938 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-catalog-content\") pod \"community-operators-cl7r6\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.470023 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-utilities\") pod \"community-operators-cl7r6\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.578443 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-utilities\") pod \"community-operators-cl7r6\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.578693 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfrgm\" (UniqueName: \"kubernetes.io/projected/a36702b9-dcfa-4945-80ac-01ac96b0e375-kube-api-access-sfrgm\") pod \"community-operators-cl7r6\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.578773 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-catalog-content\") pod \"community-operators-cl7r6\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.579095 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-utilities\") pod \"community-operators-cl7r6\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.579477 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-catalog-content\") pod \"community-operators-cl7r6\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.614824 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfrgm\" (UniqueName: \"kubernetes.io/projected/a36702b9-dcfa-4945-80ac-01ac96b0e375-kube-api-access-sfrgm\") pod \"community-operators-cl7r6\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:24 crc kubenswrapper[4751]: I0316 00:21:24.685771 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:27 crc kubenswrapper[4751]: I0316 00:21:27.854737 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:21:27 crc kubenswrapper[4751]: I0316 00:21:27.855831 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:21:29 crc kubenswrapper[4751]: E0316 00:21:29.833546 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/smart-gateway-operator:latest" Mar 16 00:21:29 crc kubenswrapper[4751]: E0316 00:21:29.833781 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/smart-gateway-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:smart-gateway-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:ANSIBLE_VERBOSITY_SMARTGATEWAY_SMARTGATEWAY_INFRA_WATCH,Value:4,ValueFrom:nil,},EnvVar{Name:ANSIBLE_DEBUG_LOGS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CORE_SMARTGATEWAY_IMAGE,Value:quay.io/infrawatch/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BRIDGE_SMARTGATEWAY_IMAGE,Value:quay.io/infrawatch/sg-bridge:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:smart-gateway-operator.v5.0.1773615492,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qtk9c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod smart-gateway-operator-697d54f584-rdntd_service-telemetry(6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 16 00:21:29 crc kubenswrapper[4751]: E0316 00:21:29.835086 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" podUID="6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7" Mar 16 00:21:29 crc kubenswrapper[4751]: E0316 00:21:29.921862 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/service-telemetry-operator:latest" Mar 16 00:21:29 crc kubenswrapper[4751]: E0316 00:21:29.922205 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/service-telemetry-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:service-telemetry-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_WEBHOOK_SNMP_IMAGE,Value:quay.io/infrawatch/prometheus-webhook-snmp:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_IMAGE,Value:quay.io/prometheus/prometheus:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER_IMAGE,Value:quay.io/prometheus/alertmanager:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:service-telemetry-operator.v1.5.1773615493,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8wnmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod service-telemetry-operator-fcbbbcfcf-ngbch_service-telemetry(2d362a30-4cc0-43f3-bae1-482bcc3ad4e5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 16 00:21:29 crc kubenswrapper[4751]: E0316 00:21:29.931586 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" podUID="2d362a30-4cc0-43f3-bae1-482bcc3ad4e5" Mar 16 00:21:30 crc kubenswrapper[4751]: I0316 00:21:30.254030 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cl7r6"] Mar 16 00:21:30 crc kubenswrapper[4751]: W0316 00:21:30.262532 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda36702b9_dcfa_4945_80ac_01ac96b0e375.slice/crio-412cc69179e70c3f14a81ba5ace5727d222d5796c3ab0f588fafa76ccc2fb2cf WatchSource:0}: Error finding container 412cc69179e70c3f14a81ba5ace5727d222d5796c3ab0f588fafa76ccc2fb2cf: Status 404 returned error can't find the container with id 412cc69179e70c3f14a81ba5ace5727d222d5796c3ab0f588fafa76ccc2fb2cf Mar 16 00:21:30 crc kubenswrapper[4751]: I0316 00:21:30.571365 4751 generic.go:334] "Generic (PLEG): container finished" podID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerID="25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913" exitCode=0 Mar 16 00:21:30 crc kubenswrapper[4751]: I0316 00:21:30.571737 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cl7r6" event={"ID":"a36702b9-dcfa-4945-80ac-01ac96b0e375","Type":"ContainerDied","Data":"25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913"} Mar 16 00:21:30 crc kubenswrapper[4751]: I0316 00:21:30.571834 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cl7r6" event={"ID":"a36702b9-dcfa-4945-80ac-01ac96b0e375","Type":"ContainerStarted","Data":"412cc69179e70c3f14a81ba5ace5727d222d5796c3ab0f588fafa76ccc2fb2cf"} Mar 16 00:21:30 crc kubenswrapper[4751]: I0316 00:21:30.574239 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-w5bqg" event={"ID":"cdda1cb0-0c2b-44e1-b1ca-942e8da817aa","Type":"ContainerStarted","Data":"9f0fece0076550a90875cd4470526fe266e29a04a96a9446073965b0d954f178"} Mar 16 00:21:30 crc kubenswrapper[4751]: E0316 00:21:30.575870 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/service-telemetry-operator:latest\\\"\"" pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" podUID="2d362a30-4cc0-43f3-bae1-482bcc3ad4e5" Mar 16 00:21:30 crc kubenswrapper[4751]: E0316 00:21:30.576073 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/smart-gateway-operator:latest\\\"\"" pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" podUID="6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7" Mar 16 00:21:30 crc kubenswrapper[4751]: I0316 00:21:30.650147 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-w5bqg" podStartSLOduration=14.170272494 podStartE2EDuration="26.650059495s" podCreationTimestamp="2026-03-16 00:21:04 +0000 UTC" firstStartedPulling="2026-03-16 00:21:05.3394663 +0000 UTC m=+893.608798590" lastFinishedPulling="2026-03-16 00:21:17.819253301 +0000 UTC m=+906.088585591" observedRunningTime="2026-03-16 00:21:30.646050812 +0000 UTC m=+918.915383112" watchObservedRunningTime="2026-03-16 00:21:30.650059495 +0000 UTC m=+918.919391795" Mar 16 00:21:31 crc kubenswrapper[4751]: I0316 00:21:31.584906 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cl7r6" event={"ID":"a36702b9-dcfa-4945-80ac-01ac96b0e375","Type":"ContainerStarted","Data":"cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397"} Mar 16 00:21:32 crc kubenswrapper[4751]: I0316 00:21:32.607768 4751 generic.go:334] "Generic (PLEG): container finished" podID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerID="cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397" exitCode=0 Mar 16 00:21:32 crc kubenswrapper[4751]: I0316 00:21:32.607816 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cl7r6" event={"ID":"a36702b9-dcfa-4945-80ac-01ac96b0e375","Type":"ContainerDied","Data":"cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397"} Mar 16 00:21:33 crc kubenswrapper[4751]: I0316 00:21:33.619558 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cl7r6" event={"ID":"a36702b9-dcfa-4945-80ac-01ac96b0e375","Type":"ContainerStarted","Data":"337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322"} Mar 16 00:21:33 crc kubenswrapper[4751]: I0316 00:21:33.649043 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cl7r6" podStartSLOduration=7.213754617 podStartE2EDuration="9.649012614s" podCreationTimestamp="2026-03-16 00:21:24 +0000 UTC" firstStartedPulling="2026-03-16 00:21:30.572825554 +0000 UTC m=+918.842157844" lastFinishedPulling="2026-03-16 00:21:33.008083541 +0000 UTC m=+921.277415841" observedRunningTime="2026-03-16 00:21:33.645934837 +0000 UTC m=+921.915267177" watchObservedRunningTime="2026-03-16 00:21:33.649012614 +0000 UTC m=+921.918344944" Mar 16 00:21:34 crc kubenswrapper[4751]: I0316 00:21:34.685938 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:34 crc kubenswrapper[4751]: I0316 00:21:34.686071 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:35 crc kubenswrapper[4751]: I0316 00:21:35.753805 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-cl7r6" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerName="registry-server" probeResult="failure" output=< Mar 16 00:21:35 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Mar 16 00:21:35 crc kubenswrapper[4751]: > Mar 16 00:21:44 crc kubenswrapper[4751]: I0316 00:21:44.705265 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" event={"ID":"2d362a30-4cc0-43f3-bae1-482bcc3ad4e5","Type":"ContainerStarted","Data":"224821b4abafd690f252487129c15595c2a9a979713cb1464922516d10d19c54"} Mar 16 00:21:44 crc kubenswrapper[4751]: I0316 00:21:44.740362 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-fcbbbcfcf-ngbch" podStartSLOduration=1.8717699209999998 podStartE2EDuration="43.740326421s" podCreationTimestamp="2026-03-16 00:21:01 +0000 UTC" firstStartedPulling="2026-03-16 00:21:02.176693217 +0000 UTC m=+890.446025507" lastFinishedPulling="2026-03-16 00:21:44.045249717 +0000 UTC m=+932.314582007" observedRunningTime="2026-03-16 00:21:44.727380117 +0000 UTC m=+932.996712437" watchObservedRunningTime="2026-03-16 00:21:44.740326421 +0000 UTC m=+933.009658751" Mar 16 00:21:44 crc kubenswrapper[4751]: I0316 00:21:44.758580 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:44 crc kubenswrapper[4751]: I0316 00:21:44.812889 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:45 crc kubenswrapper[4751]: I0316 00:21:45.715616 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" event={"ID":"6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7","Type":"ContainerStarted","Data":"78ed7411b6d0d0c8d6ab285d5e6683e67a5534c079df734c78c88ec7bed546fc"} Mar 16 00:21:45 crc kubenswrapper[4751]: I0316 00:21:45.746572 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-697d54f584-rdntd" podStartSLOduration=1.700943091 podStartE2EDuration="42.746539989s" podCreationTimestamp="2026-03-16 00:21:03 +0000 UTC" firstStartedPulling="2026-03-16 00:21:04.11449902 +0000 UTC m=+892.383831300" lastFinishedPulling="2026-03-16 00:21:45.160095898 +0000 UTC m=+933.429428198" observedRunningTime="2026-03-16 00:21:45.736650031 +0000 UTC m=+934.005982351" watchObservedRunningTime="2026-03-16 00:21:45.746539989 +0000 UTC m=+934.015872309" Mar 16 00:21:47 crc kubenswrapper[4751]: I0316 00:21:47.701964 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cl7r6"] Mar 16 00:21:47 crc kubenswrapper[4751]: I0316 00:21:47.702599 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cl7r6" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerName="registry-server" containerID="cri-o://337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322" gracePeriod=2 Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.142505 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.220275 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfrgm\" (UniqueName: \"kubernetes.io/projected/a36702b9-dcfa-4945-80ac-01ac96b0e375-kube-api-access-sfrgm\") pod \"a36702b9-dcfa-4945-80ac-01ac96b0e375\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.220329 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-utilities\") pod \"a36702b9-dcfa-4945-80ac-01ac96b0e375\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.220372 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-catalog-content\") pod \"a36702b9-dcfa-4945-80ac-01ac96b0e375\" (UID: \"a36702b9-dcfa-4945-80ac-01ac96b0e375\") " Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.221286 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-utilities" (OuterVolumeSpecName: "utilities") pod "a36702b9-dcfa-4945-80ac-01ac96b0e375" (UID: "a36702b9-dcfa-4945-80ac-01ac96b0e375"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.226631 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36702b9-dcfa-4945-80ac-01ac96b0e375-kube-api-access-sfrgm" (OuterVolumeSpecName: "kube-api-access-sfrgm") pod "a36702b9-dcfa-4945-80ac-01ac96b0e375" (UID: "a36702b9-dcfa-4945-80ac-01ac96b0e375"). InnerVolumeSpecName "kube-api-access-sfrgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.291927 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a36702b9-dcfa-4945-80ac-01ac96b0e375" (UID: "a36702b9-dcfa-4945-80ac-01ac96b0e375"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.321963 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfrgm\" (UniqueName: \"kubernetes.io/projected/a36702b9-dcfa-4945-80ac-01ac96b0e375-kube-api-access-sfrgm\") on node \"crc\" DevicePath \"\"" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.322001 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.322018 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36702b9-dcfa-4945-80ac-01ac96b0e375-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.736543 4751 generic.go:334] "Generic (PLEG): container finished" podID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerID="337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322" exitCode=0 Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.736580 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cl7r6" event={"ID":"a36702b9-dcfa-4945-80ac-01ac96b0e375","Type":"ContainerDied","Data":"337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322"} Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.736604 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cl7r6" event={"ID":"a36702b9-dcfa-4945-80ac-01ac96b0e375","Type":"ContainerDied","Data":"412cc69179e70c3f14a81ba5ace5727d222d5796c3ab0f588fafa76ccc2fb2cf"} Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.736620 4751 scope.go:117] "RemoveContainer" containerID="337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.736729 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cl7r6" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.759844 4751 scope.go:117] "RemoveContainer" containerID="cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.762924 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cl7r6"] Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.771877 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cl7r6"] Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.796406 4751 scope.go:117] "RemoveContainer" containerID="25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.817755 4751 scope.go:117] "RemoveContainer" containerID="337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322" Mar 16 00:21:48 crc kubenswrapper[4751]: E0316 00:21:48.826545 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322\": container with ID starting with 337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322 not found: ID does not exist" containerID="337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.826578 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322"} err="failed to get container status \"337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322\": rpc error: code = NotFound desc = could not find container \"337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322\": container with ID starting with 337bf90bc4dcf05cfc9984ab3f947fa9782d9650fdeceb48f533e1e94796b322 not found: ID does not exist" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.826600 4751 scope.go:117] "RemoveContainer" containerID="cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397" Mar 16 00:21:48 crc kubenswrapper[4751]: E0316 00:21:48.826821 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397\": container with ID starting with cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397 not found: ID does not exist" containerID="cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.826839 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397"} err="failed to get container status \"cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397\": rpc error: code = NotFound desc = could not find container \"cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397\": container with ID starting with cfc5e4a135a714a5a37cf39c708e7940ecc665d3d0fa64bd398f820400685397 not found: ID does not exist" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.826852 4751 scope.go:117] "RemoveContainer" containerID="25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913" Mar 16 00:21:48 crc kubenswrapper[4751]: E0316 00:21:48.827021 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913\": container with ID starting with 25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913 not found: ID does not exist" containerID="25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913" Mar 16 00:21:48 crc kubenswrapper[4751]: I0316 00:21:48.827040 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913"} err="failed to get container status \"25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913\": rpc error: code = NotFound desc = could not find container \"25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913\": container with ID starting with 25e8f6c47e63b2c9a6279775b91fa7c2917e4172e1f2c12633002f5d5e5f6913 not found: ID does not exist" Mar 16 00:21:50 crc kubenswrapper[4751]: I0316 00:21:50.465419 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" path="/var/lib/kubelet/pods/a36702b9-dcfa-4945-80ac-01ac96b0e375/volumes" Mar 16 00:21:57 crc kubenswrapper[4751]: I0316 00:21:57.854438 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:21:57 crc kubenswrapper[4751]: I0316 00:21:57.856191 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:21:57 crc kubenswrapper[4751]: I0316 00:21:57.856243 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:21:57 crc kubenswrapper[4751]: I0316 00:21:57.856760 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65efad03d473120c12153253332bfeb08a8e6a82c87fa93d5e003a10153418a4"} pod="openshift-machine-config-operator/machine-config-daemon-6fccz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 16 00:21:57 crc kubenswrapper[4751]: I0316 00:21:57.856809 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" containerID="cri-o://65efad03d473120c12153253332bfeb08a8e6a82c87fa93d5e003a10153418a4" gracePeriod=600 Mar 16 00:21:58 crc kubenswrapper[4751]: I0316 00:21:58.814138 4751 generic.go:334] "Generic (PLEG): container finished" podID="25429719-d18d-4909-8455-701d09423c36" containerID="65efad03d473120c12153253332bfeb08a8e6a82c87fa93d5e003a10153418a4" exitCode=0 Mar 16 00:21:58 crc kubenswrapper[4751]: I0316 00:21:58.814323 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerDied","Data":"65efad03d473120c12153253332bfeb08a8e6a82c87fa93d5e003a10153418a4"} Mar 16 00:21:58 crc kubenswrapper[4751]: I0316 00:21:58.814576 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerStarted","Data":"8e0586c280086ba5e014ebfa12b8cdaadcbcd7999bb69b5913e321ce4b81f9a1"} Mar 16 00:21:58 crc kubenswrapper[4751]: I0316 00:21:58.814596 4751 scope.go:117] "RemoveContainer" containerID="a45820bd88f0ba2e569a216d14a8ef916b60532a9717b18faa5c4a39cc910750" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.147363 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560342-qpmsc"] Mar 16 00:22:00 crc kubenswrapper[4751]: E0316 00:22:00.147988 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerName="extract-content" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.148010 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerName="extract-content" Mar 16 00:22:00 crc kubenswrapper[4751]: E0316 00:22:00.148042 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerName="extract-utilities" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.148055 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerName="extract-utilities" Mar 16 00:22:00 crc kubenswrapper[4751]: E0316 00:22:00.148073 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerName="registry-server" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.148086 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerName="registry-server" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.148295 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36702b9-dcfa-4945-80ac-01ac96b0e375" containerName="registry-server" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.148866 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560342-qpmsc" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.154328 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.154668 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.155515 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.193364 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560342-qpmsc"] Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.206889 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvb77\" (UniqueName: \"kubernetes.io/projected/6dcc0fca-049b-4dbe-9470-8e4b89bccfb6-kube-api-access-gvb77\") pod \"auto-csr-approver-29560342-qpmsc\" (UID: \"6dcc0fca-049b-4dbe-9470-8e4b89bccfb6\") " pod="openshift-infra/auto-csr-approver-29560342-qpmsc" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.309190 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvb77\" (UniqueName: \"kubernetes.io/projected/6dcc0fca-049b-4dbe-9470-8e4b89bccfb6-kube-api-access-gvb77\") pod \"auto-csr-approver-29560342-qpmsc\" (UID: \"6dcc0fca-049b-4dbe-9470-8e4b89bccfb6\") " pod="openshift-infra/auto-csr-approver-29560342-qpmsc" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.349390 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvb77\" (UniqueName: \"kubernetes.io/projected/6dcc0fca-049b-4dbe-9470-8e4b89bccfb6-kube-api-access-gvb77\") pod \"auto-csr-approver-29560342-qpmsc\" (UID: \"6dcc0fca-049b-4dbe-9470-8e4b89bccfb6\") " pod="openshift-infra/auto-csr-approver-29560342-qpmsc" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.467325 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560342-qpmsc" Mar 16 00:22:00 crc kubenswrapper[4751]: I0316 00:22:00.884339 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560342-qpmsc"] Mar 16 00:22:00 crc kubenswrapper[4751]: W0316 00:22:00.890545 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dcc0fca_049b_4dbe_9470_8e4b89bccfb6.slice/crio-bad83de2929be003a481ca496fdd4d2a285b3fbfd5e6e9d7b5662cd8f6a955d7 WatchSource:0}: Error finding container bad83de2929be003a481ca496fdd4d2a285b3fbfd5e6e9d7b5662cd8f6a955d7: Status 404 returned error can't find the container with id bad83de2929be003a481ca496fdd4d2a285b3fbfd5e6e9d7b5662cd8f6a955d7 Mar 16 00:22:01 crc kubenswrapper[4751]: I0316 00:22:01.840293 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560342-qpmsc" event={"ID":"6dcc0fca-049b-4dbe-9470-8e4b89bccfb6","Type":"ContainerStarted","Data":"bad83de2929be003a481ca496fdd4d2a285b3fbfd5e6e9d7b5662cd8f6a955d7"} Mar 16 00:22:02 crc kubenswrapper[4751]: I0316 00:22:02.848065 4751 generic.go:334] "Generic (PLEG): container finished" podID="6dcc0fca-049b-4dbe-9470-8e4b89bccfb6" containerID="e8a67af31e0d598a8cc3f6536d0d51d0a1e24bddb2370c1a6cc411cb39bc04e8" exitCode=0 Mar 16 00:22:02 crc kubenswrapper[4751]: I0316 00:22:02.848146 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560342-qpmsc" event={"ID":"6dcc0fca-049b-4dbe-9470-8e4b89bccfb6","Type":"ContainerDied","Data":"e8a67af31e0d598a8cc3f6536d0d51d0a1e24bddb2370c1a6cc411cb39bc04e8"} Mar 16 00:22:04 crc kubenswrapper[4751]: I0316 00:22:04.184355 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560342-qpmsc" Mar 16 00:22:04 crc kubenswrapper[4751]: I0316 00:22:04.257231 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvb77\" (UniqueName: \"kubernetes.io/projected/6dcc0fca-049b-4dbe-9470-8e4b89bccfb6-kube-api-access-gvb77\") pod \"6dcc0fca-049b-4dbe-9470-8e4b89bccfb6\" (UID: \"6dcc0fca-049b-4dbe-9470-8e4b89bccfb6\") " Mar 16 00:22:04 crc kubenswrapper[4751]: I0316 00:22:04.262476 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dcc0fca-049b-4dbe-9470-8e4b89bccfb6-kube-api-access-gvb77" (OuterVolumeSpecName: "kube-api-access-gvb77") pod "6dcc0fca-049b-4dbe-9470-8e4b89bccfb6" (UID: "6dcc0fca-049b-4dbe-9470-8e4b89bccfb6"). InnerVolumeSpecName "kube-api-access-gvb77". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:22:04 crc kubenswrapper[4751]: I0316 00:22:04.358621 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvb77\" (UniqueName: \"kubernetes.io/projected/6dcc0fca-049b-4dbe-9470-8e4b89bccfb6-kube-api-access-gvb77\") on node \"crc\" DevicePath \"\"" Mar 16 00:22:04 crc kubenswrapper[4751]: I0316 00:22:04.869224 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560342-qpmsc" event={"ID":"6dcc0fca-049b-4dbe-9470-8e4b89bccfb6","Type":"ContainerDied","Data":"bad83de2929be003a481ca496fdd4d2a285b3fbfd5e6e9d7b5662cd8f6a955d7"} Mar 16 00:22:04 crc kubenswrapper[4751]: I0316 00:22:04.869296 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bad83de2929be003a481ca496fdd4d2a285b3fbfd5e6e9d7b5662cd8f6a955d7" Mar 16 00:22:04 crc kubenswrapper[4751]: I0316 00:22:04.869309 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560342-qpmsc" Mar 16 00:22:05 crc kubenswrapper[4751]: I0316 00:22:05.296171 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29560336-pvk5d"] Mar 16 00:22:05 crc kubenswrapper[4751]: I0316 00:22:05.309678 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29560336-pvk5d"] Mar 16 00:22:06 crc kubenswrapper[4751]: I0316 00:22:06.468710 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9713fca-2f31-4266-809a-76c2085543ae" path="/var/lib/kubelet/pods/d9713fca-2f31-4266-809a-76c2085543ae/volumes" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.055031 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xtrt6"] Mar 16 00:22:13 crc kubenswrapper[4751]: E0316 00:22:13.055833 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dcc0fca-049b-4dbe-9470-8e4b89bccfb6" containerName="oc" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.055848 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dcc0fca-049b-4dbe-9470-8e4b89bccfb6" containerName="oc" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.055992 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dcc0fca-049b-4dbe-9470-8e4b89bccfb6" containerName="oc" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.056451 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.060151 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.060283 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-z9qm4" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.060471 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.062129 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.062518 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.062619 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.062725 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.070716 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xtrt6"] Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.181523 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.181574 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-config\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.181598 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.181616 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkkmz\" (UniqueName: \"kubernetes.io/projected/58867d99-7102-43c5-9e18-eb59e2c702b6-kube-api-access-kkkmz\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.181744 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.181793 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-users\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.181838 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.283301 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.283376 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkkmz\" (UniqueName: \"kubernetes.io/projected/58867d99-7102-43c5-9e18-eb59e2c702b6-kube-api-access-kkkmz\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.283437 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.284147 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-users\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.284205 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.284277 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.284307 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-config\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.285322 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-config\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.288726 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-users\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.288787 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.289632 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.296503 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.300402 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkkmz\" (UniqueName: \"kubernetes.io/projected/58867d99-7102-43c5-9e18-eb59e2c702b6-kube-api-access-kkkmz\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.301754 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-xtrt6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.372634 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.612610 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xtrt6"] Mar 16 00:22:13 crc kubenswrapper[4751]: I0316 00:22:13.926372 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" event={"ID":"58867d99-7102-43c5-9e18-eb59e2c702b6","Type":"ContainerStarted","Data":"2bcc8dec9dada6e5bc5b7049070c865bf6fc0782debf5eb75c3e652bb1448dac"} Mar 16 00:22:17 crc kubenswrapper[4751]: I0316 00:22:17.660165 4751 scope.go:117] "RemoveContainer" containerID="fcec6157b1ec83f0230be70b6805d656ab0b1106c392804f4eb26cffe0f43df3" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.925224 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.951446 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.953824 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-session-secret" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.954216 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.954863 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-2" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.954970 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-tls-assets-0" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.955007 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"serving-certs-ca-bundle" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.955246 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-0" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.955363 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-prometheus-proxy-tls" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.955504 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-web-config" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.955528 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-1" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.957281 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-stf-dockercfg-24qnp" Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.960451 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 16 00:22:22 crc kubenswrapper[4751]: I0316 00:22:22.991233 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" event={"ID":"58867d99-7102-43c5-9e18-eb59e2c702b6","Type":"ContainerStarted","Data":"fbc2394f9179ee1cc977dd5f4fa92be4efab1cbb916ee226b3a509e55797b5f1"} Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.020574 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-24aee5b7-8de2-4b8f-a7f7-67b4ca950e34\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-24aee5b7-8de2-4b8f-a7f7-67b4ca950e34\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021001 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021078 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc7wb\" (UniqueName: \"kubernetes.io/projected/aaf54795-c77b-42bf-a088-5104d905ab2d-kube-api-access-tc7wb\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021176 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-config\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021257 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-web-config\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021350 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021431 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021527 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aaf54795-c77b-42bf-a088-5104d905ab2d-config-out\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021629 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021669 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021717 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.021751 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aaf54795-c77b-42bf-a088-5104d905ab2d-tls-assets\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123325 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123380 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aaf54795-c77b-42bf-a088-5104d905ab2d-tls-assets\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123430 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-24aee5b7-8de2-4b8f-a7f7-67b4ca950e34\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-24aee5b7-8de2-4b8f-a7f7-67b4ca950e34\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123453 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123481 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc7wb\" (UniqueName: \"kubernetes.io/projected/aaf54795-c77b-42bf-a088-5104d905ab2d-kube-api-access-tc7wb\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123510 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-config\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123532 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-web-config\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123629 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123680 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123705 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aaf54795-c77b-42bf-a088-5104d905ab2d-config-out\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123744 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.123768 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.125607 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: E0316 00:22:23.125804 4751 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 16 00:22:23 crc kubenswrapper[4751]: E0316 00:22:23.125902 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-prometheus-proxy-tls podName:aaf54795-c77b-42bf-a088-5104d905ab2d nodeName:}" failed. No retries permitted until 2026-03-16 00:22:23.625875126 +0000 UTC m=+971.895207456 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "aaf54795-c77b-42bf-a088-5104d905ab2d") : secret "default-prometheus-proxy-tls" not found Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.127548 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.127610 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.127619 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaf54795-c77b-42bf-a088-5104d905ab2d-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.132424 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-config\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.132831 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aaf54795-c77b-42bf-a088-5104d905ab2d-config-out\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.133418 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-web-config\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.134789 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.135006 4751 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.135093 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-24aee5b7-8de2-4b8f-a7f7-67b4ca950e34\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-24aee5b7-8de2-4b8f-a7f7-67b4ca950e34\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4986bc2c2a6afd3b77e2d3971be27af5a742b68bb0a09b09a7adb4578a2da16d/globalmount\"" pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.143964 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aaf54795-c77b-42bf-a088-5104d905ab2d-tls-assets\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.149481 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc7wb\" (UniqueName: \"kubernetes.io/projected/aaf54795-c77b-42bf-a088-5104d905ab2d-kube-api-access-tc7wb\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.158988 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-24aee5b7-8de2-4b8f-a7f7-67b4ca950e34\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-24aee5b7-8de2-4b8f-a7f7-67b4ca950e34\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: I0316 00:22:23.631762 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:23 crc kubenswrapper[4751]: E0316 00:22:23.631888 4751 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 16 00:22:23 crc kubenswrapper[4751]: E0316 00:22:23.631939 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-prometheus-proxy-tls podName:aaf54795-c77b-42bf-a088-5104d905ab2d nodeName:}" failed. No retries permitted until 2026-03-16 00:22:24.631926877 +0000 UTC m=+972.901259167 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "aaf54795-c77b-42bf-a088-5104d905ab2d") : secret "default-prometheus-proxy-tls" not found Mar 16 00:22:24 crc kubenswrapper[4751]: I0316 00:22:24.646912 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:24 crc kubenswrapper[4751]: I0316 00:22:24.655396 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/aaf54795-c77b-42bf-a088-5104d905ab2d-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"aaf54795-c77b-42bf-a088-5104d905ab2d\") " pod="service-telemetry/prometheus-default-0" Mar 16 00:22:24 crc kubenswrapper[4751]: I0316 00:22:24.769446 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 16 00:22:25 crc kubenswrapper[4751]: I0316 00:22:25.249152 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" podStartSLOduration=3.216462012 podStartE2EDuration="12.249132816s" podCreationTimestamp="2026-03-16 00:22:13 +0000 UTC" firstStartedPulling="2026-03-16 00:22:13.627760181 +0000 UTC m=+961.897092471" lastFinishedPulling="2026-03-16 00:22:22.660430975 +0000 UTC m=+970.929763275" observedRunningTime="2026-03-16 00:22:23.008932889 +0000 UTC m=+971.278265169" watchObservedRunningTime="2026-03-16 00:22:25.249132816 +0000 UTC m=+973.518465106" Mar 16 00:22:25 crc kubenswrapper[4751]: I0316 00:22:25.253597 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 16 00:22:26 crc kubenswrapper[4751]: I0316 00:22:26.019158 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"aaf54795-c77b-42bf-a088-5104d905ab2d","Type":"ContainerStarted","Data":"b257fc1d53f41a80ce98fab999b1b29994ba083dc9d3e632849026cfb79c8538"} Mar 16 00:22:30 crc kubenswrapper[4751]: I0316 00:22:30.054431 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"aaf54795-c77b-42bf-a088-5104d905ab2d","Type":"ContainerStarted","Data":"c7c73ca9cdd7258bda115011bacbea284e10c165f558e8b6faceba54d160ed05"} Mar 16 00:22:32 crc kubenswrapper[4751]: I0316 00:22:32.702876 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n"] Mar 16 00:22:32 crc kubenswrapper[4751]: I0316 00:22:32.704567 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n" Mar 16 00:22:32 crc kubenswrapper[4751]: I0316 00:22:32.717172 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n"] Mar 16 00:22:32 crc kubenswrapper[4751]: I0316 00:22:32.774361 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6zl5\" (UniqueName: \"kubernetes.io/projected/ec8ad70c-3602-46da-8cd2-b344f5365a71-kube-api-access-z6zl5\") pod \"default-snmp-webhook-78bcbbdcff-pg68n\" (UID: \"ec8ad70c-3602-46da-8cd2-b344f5365a71\") " pod="service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n" Mar 16 00:22:32 crc kubenswrapper[4751]: I0316 00:22:32.876630 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6zl5\" (UniqueName: \"kubernetes.io/projected/ec8ad70c-3602-46da-8cd2-b344f5365a71-kube-api-access-z6zl5\") pod \"default-snmp-webhook-78bcbbdcff-pg68n\" (UID: \"ec8ad70c-3602-46da-8cd2-b344f5365a71\") " pod="service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n" Mar 16 00:22:32 crc kubenswrapper[4751]: I0316 00:22:32.897774 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6zl5\" (UniqueName: \"kubernetes.io/projected/ec8ad70c-3602-46da-8cd2-b344f5365a71-kube-api-access-z6zl5\") pod \"default-snmp-webhook-78bcbbdcff-pg68n\" (UID: \"ec8ad70c-3602-46da-8cd2-b344f5365a71\") " pod="service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n" Mar 16 00:22:33 crc kubenswrapper[4751]: I0316 00:22:33.038289 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n" Mar 16 00:22:33 crc kubenswrapper[4751]: I0316 00:22:33.508559 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n"] Mar 16 00:22:34 crc kubenswrapper[4751]: I0316 00:22:34.084310 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n" event={"ID":"ec8ad70c-3602-46da-8cd2-b344f5365a71","Type":"ContainerStarted","Data":"effed91519e589a3168116bda3f83d61cc5ebde0e383b4ceb4ec4c23f3efd567"} Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.344290 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.349020 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.353627 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-cluster-tls-config" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.353678 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-tls-assets-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.353740 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-web-config" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.353844 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-generated" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.353738 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-alertmanager-proxy-tls" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.354064 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-stf-dockercfg-wdfxz" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.357812 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.446779 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.446882 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-config-volume\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.446911 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.446964 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.447028 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/32d5f19b-e735-49bd-8212-3367fd97cc49-tls-assets\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.447066 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-web-config\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.447376 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/32d5f19b-e735-49bd-8212-3367fd97cc49-config-out\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.447426 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zttxg\" (UniqueName: \"kubernetes.io/projected/32d5f19b-e735-49bd-8212-3367fd97cc49-kube-api-access-zttxg\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.447470 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a3c547b7-1a71-4462-a808-ca00ac7df1d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3c547b7-1a71-4462-a808-ca00ac7df1d6\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.548447 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.548513 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-config-volume\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.548542 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.548579 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.548622 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/32d5f19b-e735-49bd-8212-3367fd97cc49-tls-assets\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.548660 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-web-config\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.548755 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zttxg\" (UniqueName: \"kubernetes.io/projected/32d5f19b-e735-49bd-8212-3367fd97cc49-kube-api-access-zttxg\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: E0316 00:22:36.548762 4751 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.548779 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/32d5f19b-e735-49bd-8212-3367fd97cc49-config-out\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.548808 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a3c547b7-1a71-4462-a808-ca00ac7df1d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3c547b7-1a71-4462-a808-ca00ac7df1d6\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: E0316 00:22:36.548834 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls podName:32d5f19b-e735-49bd-8212-3367fd97cc49 nodeName:}" failed. No retries permitted until 2026-03-16 00:22:37.04881417 +0000 UTC m=+985.318146480 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "32d5f19b-e735-49bd-8212-3367fd97cc49") : secret "default-alertmanager-proxy-tls" not found Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.555934 4751 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.556023 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a3c547b7-1a71-4462-a808-ca00ac7df1d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3c547b7-1a71-4462-a808-ca00ac7df1d6\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e8d158049d4e6e7dcb01bc395021d1062b14ff4366d9530bd5af7c828305bd73/globalmount\"" pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.557754 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/32d5f19b-e735-49bd-8212-3367fd97cc49-config-out\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.561519 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-web-config\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.562582 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/32d5f19b-e735-49bd-8212-3367fd97cc49-tls-assets\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.563231 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.566036 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.571730 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-config-volume\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.574052 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zttxg\" (UniqueName: \"kubernetes.io/projected/32d5f19b-e735-49bd-8212-3367fd97cc49-kube-api-access-zttxg\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:36 crc kubenswrapper[4751]: I0316 00:22:36.592642 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a3c547b7-1a71-4462-a808-ca00ac7df1d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3c547b7-1a71-4462-a808-ca00ac7df1d6\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:37 crc kubenswrapper[4751]: I0316 00:22:37.055666 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:37 crc kubenswrapper[4751]: E0316 00:22:37.056181 4751 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 16 00:22:37 crc kubenswrapper[4751]: E0316 00:22:37.056249 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls podName:32d5f19b-e735-49bd-8212-3367fd97cc49 nodeName:}" failed. No retries permitted until 2026-03-16 00:22:38.05622714 +0000 UTC m=+986.325559450 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "32d5f19b-e735-49bd-8212-3367fd97cc49") : secret "default-alertmanager-proxy-tls" not found Mar 16 00:22:38 crc kubenswrapper[4751]: I0316 00:22:38.069288 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:38 crc kubenswrapper[4751]: E0316 00:22:38.069476 4751 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 16 00:22:38 crc kubenswrapper[4751]: E0316 00:22:38.069575 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls podName:32d5f19b-e735-49bd-8212-3367fd97cc49 nodeName:}" failed. No retries permitted until 2026-03-16 00:22:40.069553037 +0000 UTC m=+988.338885337 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "32d5f19b-e735-49bd-8212-3367fd97cc49") : secret "default-alertmanager-proxy-tls" not found Mar 16 00:22:39 crc kubenswrapper[4751]: I0316 00:22:39.149340 4751 generic.go:334] "Generic (PLEG): container finished" podID="aaf54795-c77b-42bf-a088-5104d905ab2d" containerID="c7c73ca9cdd7258bda115011bacbea284e10c165f558e8b6faceba54d160ed05" exitCode=0 Mar 16 00:22:39 crc kubenswrapper[4751]: I0316 00:22:39.149446 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"aaf54795-c77b-42bf-a088-5104d905ab2d","Type":"ContainerDied","Data":"c7c73ca9cdd7258bda115011bacbea284e10c165f558e8b6faceba54d160ed05"} Mar 16 00:22:40 crc kubenswrapper[4751]: I0316 00:22:40.095278 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:40 crc kubenswrapper[4751]: I0316 00:22:40.119323 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/32d5f19b-e735-49bd-8212-3367fd97cc49-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"32d5f19b-e735-49bd-8212-3367fd97cc49\") " pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:40 crc kubenswrapper[4751]: I0316 00:22:40.272347 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 16 00:22:42 crc kubenswrapper[4751]: I0316 00:22:42.171789 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n" event={"ID":"ec8ad70c-3602-46da-8cd2-b344f5365a71","Type":"ContainerStarted","Data":"2e6f33b5ea9c77326d20a2420ffc6d02cbc8f6aa238fba7247721e112ab2d70c"} Mar 16 00:22:42 crc kubenswrapper[4751]: I0316 00:22:42.188971 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-pg68n" podStartSLOduration=1.809069605 podStartE2EDuration="10.188949485s" podCreationTimestamp="2026-03-16 00:22:32 +0000 UTC" firstStartedPulling="2026-03-16 00:22:33.515767082 +0000 UTC m=+981.785099372" lastFinishedPulling="2026-03-16 00:22:41.895646962 +0000 UTC m=+990.164979252" observedRunningTime="2026-03-16 00:22:42.186614509 +0000 UTC m=+990.455946819" watchObservedRunningTime="2026-03-16 00:22:42.188949485 +0000 UTC m=+990.458281775" Mar 16 00:22:42 crc kubenswrapper[4751]: I0316 00:22:42.202562 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 16 00:22:42 crc kubenswrapper[4751]: W0316 00:22:42.335749 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32d5f19b_e735_49bd_8212_3367fd97cc49.slice/crio-094420a1ff182a3e3036393fef98d6f2930b2bd6c31c398071db280c0bbfd6cb WatchSource:0}: Error finding container 094420a1ff182a3e3036393fef98d6f2930b2bd6c31c398071db280c0bbfd6cb: Status 404 returned error can't find the container with id 094420a1ff182a3e3036393fef98d6f2930b2bd6c31c398071db280c0bbfd6cb Mar 16 00:22:43 crc kubenswrapper[4751]: I0316 00:22:43.180471 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"32d5f19b-e735-49bd-8212-3367fd97cc49","Type":"ContainerStarted","Data":"094420a1ff182a3e3036393fef98d6f2930b2bd6c31c398071db280c0bbfd6cb"} Mar 16 00:22:46 crc kubenswrapper[4751]: I0316 00:22:46.199389 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"aaf54795-c77b-42bf-a088-5104d905ab2d","Type":"ContainerStarted","Data":"e1c3213b29c27bcbfe6b492acd635a554ea4601760cdd2825b4bcc2c04a2f0a7"} Mar 16 00:22:48 crc kubenswrapper[4751]: I0316 00:22:48.213758 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"aaf54795-c77b-42bf-a088-5104d905ab2d","Type":"ContainerStarted","Data":"c5751ab9f769d927ba470ec1e1748d1a636d5397ab6ac9c6dc07a7d80c5faf53"} Mar 16 00:22:48 crc kubenswrapper[4751]: I0316 00:22:48.216283 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"32d5f19b-e735-49bd-8212-3367fd97cc49","Type":"ContainerStarted","Data":"7a1b7f8df8e635ade8918b990ea7ded1088cb7a88693ca480728d2130b78de63"} Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.043904 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq"] Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.045279 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.048776 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-meter-sg-core-configmap" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.048927 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-coll-meter-proxy-tls" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.049483 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-session-secret" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.049764 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-dockercfg-qtw26" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.059073 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/93427807-2b70-4cbe-9e6e-6fc6486a0020-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.059126 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.059147 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/93427807-2b70-4cbe-9e6e-6fc6486a0020-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.059172 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.059209 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvd4h\" (UniqueName: \"kubernetes.io/projected/93427807-2b70-4cbe-9e6e-6fc6486a0020-kube-api-access-lvd4h\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.065397 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq"] Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.160366 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/93427807-2b70-4cbe-9e6e-6fc6486a0020-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.160406 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.160448 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/93427807-2b70-4cbe-9e6e-6fc6486a0020-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.160472 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.160504 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvd4h\" (UniqueName: \"kubernetes.io/projected/93427807-2b70-4cbe-9e6e-6fc6486a0020-kube-api-access-lvd4h\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: E0316 00:22:49.160599 4751 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 16 00:22:49 crc kubenswrapper[4751]: E0316 00:22:49.160676 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-default-cloud1-coll-meter-proxy-tls podName:93427807-2b70-4cbe-9e6e-6fc6486a0020 nodeName:}" failed. No retries permitted until 2026-03-16 00:22:49.660658099 +0000 UTC m=+997.929990379 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" (UID: "93427807-2b70-4cbe-9e6e-6fc6486a0020") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.160760 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/93427807-2b70-4cbe-9e6e-6fc6486a0020-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.161470 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/93427807-2b70-4cbe-9e6e-6fc6486a0020-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.176807 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.177472 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvd4h\" (UniqueName: \"kubernetes.io/projected/93427807-2b70-4cbe-9e6e-6fc6486a0020-kube-api-access-lvd4h\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: I0316 00:22:49.668394 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:49 crc kubenswrapper[4751]: E0316 00:22:49.668577 4751 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 16 00:22:49 crc kubenswrapper[4751]: E0316 00:22:49.668659 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-default-cloud1-coll-meter-proxy-tls podName:93427807-2b70-4cbe-9e6e-6fc6486a0020 nodeName:}" failed. No retries permitted until 2026-03-16 00:22:50.668638705 +0000 UTC m=+998.937970995 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" (UID: "93427807-2b70-4cbe-9e6e-6fc6486a0020") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 16 00:22:50 crc kubenswrapper[4751]: I0316 00:22:50.684318 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:50 crc kubenswrapper[4751]: I0316 00:22:50.711004 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/93427807-2b70-4cbe-9e6e-6fc6486a0020-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq\" (UID: \"93427807-2b70-4cbe-9e6e-6fc6486a0020\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:50 crc kubenswrapper[4751]: I0316 00:22:50.866081 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.757173 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk"] Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.758347 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.762758 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-meter-sg-core-configmap" Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.762986 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-ceil-meter-proxy-tls" Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.768744 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk"] Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.899778 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.900342 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.900397 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs8pv\" (UniqueName: \"kubernetes.io/projected/ccf828f2-8f37-48da-8f95-b0e0ce744f34-kube-api-access-rs8pv\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.900420 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/ccf828f2-8f37-48da-8f95-b0e0ce744f34-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:51 crc kubenswrapper[4751]: I0316 00:22:51.900549 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ccf828f2-8f37-48da-8f95-b0e0ce744f34-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.001993 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ccf828f2-8f37-48da-8f95-b0e0ce744f34-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.002065 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.002142 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.002172 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs8pv\" (UniqueName: \"kubernetes.io/projected/ccf828f2-8f37-48da-8f95-b0e0ce744f34-kube-api-access-rs8pv\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: E0316 00:22:52.002226 4751 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 16 00:22:52 crc kubenswrapper[4751]: E0316 00:22:52.002297 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-default-cloud1-ceil-meter-proxy-tls podName:ccf828f2-8f37-48da-8f95-b0e0ce744f34 nodeName:}" failed. No retries permitted until 2026-03-16 00:22:52.502279747 +0000 UTC m=+1000.771612037 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" (UID: "ccf828f2-8f37-48da-8f95-b0e0ce744f34") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.002407 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ccf828f2-8f37-48da-8f95-b0e0ce744f34-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.002671 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/ccf828f2-8f37-48da-8f95-b0e0ce744f34-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.003490 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/ccf828f2-8f37-48da-8f95-b0e0ce744f34-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.007131 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.019298 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs8pv\" (UniqueName: \"kubernetes.io/projected/ccf828f2-8f37-48da-8f95-b0e0ce744f34-kube-api-access-rs8pv\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: I0316 00:22:52.507833 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:52 crc kubenswrapper[4751]: E0316 00:22:52.508301 4751 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 16 00:22:52 crc kubenswrapper[4751]: E0316 00:22:52.508439 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-default-cloud1-ceil-meter-proxy-tls podName:ccf828f2-8f37-48da-8f95-b0e0ce744f34 nodeName:}" failed. No retries permitted until 2026-03-16 00:22:53.508333949 +0000 UTC m=+1001.777666239 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" (UID: "ccf828f2-8f37-48da-8f95-b0e0ce744f34") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 16 00:22:53 crc kubenswrapper[4751]: I0316 00:22:53.520380 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:53 crc kubenswrapper[4751]: I0316 00:22:53.532762 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ccf828f2-8f37-48da-8f95-b0e0ce744f34-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk\" (UID: \"ccf828f2-8f37-48da-8f95-b0e0ce744f34\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:53 crc kubenswrapper[4751]: I0316 00:22:53.586491 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.376319 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp"] Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.380290 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.383387 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-sens-meter-sg-core-configmap" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.383643 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-sens-meter-proxy-tls" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.396622 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp"] Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.447724 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.447779 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/24d89dab-af8f-4823-b6a4-37e9fa41a466-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.447802 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/24d89dab-af8f-4823-b6a4-37e9fa41a466-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.447840 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q8wd\" (UniqueName: \"kubernetes.io/projected/24d89dab-af8f-4823-b6a4-37e9fa41a466-kube-api-access-2q8wd\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.447896 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.548648 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.548702 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/24d89dab-af8f-4823-b6a4-37e9fa41a466-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.548729 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/24d89dab-af8f-4823-b6a4-37e9fa41a466-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.548771 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q8wd\" (UniqueName: \"kubernetes.io/projected/24d89dab-af8f-4823-b6a4-37e9fa41a466-kube-api-access-2q8wd\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: E0316 00:22:55.548797 4751 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.548823 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: E0316 00:22:55.548865 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-default-cloud1-sens-meter-proxy-tls podName:24d89dab-af8f-4823-b6a4-37e9fa41a466 nodeName:}" failed. No retries permitted until 2026-03-16 00:22:56.048847746 +0000 UTC m=+1004.318180036 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" (UID: "24d89dab-af8f-4823-b6a4-37e9fa41a466") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.549230 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/24d89dab-af8f-4823-b6a4-37e9fa41a466-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.550043 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/24d89dab-af8f-4823-b6a4-37e9fa41a466-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.555401 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:55 crc kubenswrapper[4751]: I0316 00:22:55.568659 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q8wd\" (UniqueName: \"kubernetes.io/projected/24d89dab-af8f-4823-b6a4-37e9fa41a466-kube-api-access-2q8wd\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:56 crc kubenswrapper[4751]: I0316 00:22:56.054857 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:56 crc kubenswrapper[4751]: E0316 00:22:56.055094 4751 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 16 00:22:56 crc kubenswrapper[4751]: E0316 00:22:56.055225 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-default-cloud1-sens-meter-proxy-tls podName:24d89dab-af8f-4823-b6a4-37e9fa41a466 nodeName:}" failed. No retries permitted until 2026-03-16 00:22:57.055197266 +0000 UTC m=+1005.324529596 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" (UID: "24d89dab-af8f-4823-b6a4-37e9fa41a466") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 16 00:22:56 crc kubenswrapper[4751]: I0316 00:22:56.271786 4751 generic.go:334] "Generic (PLEG): container finished" podID="32d5f19b-e735-49bd-8212-3367fd97cc49" containerID="7a1b7f8df8e635ade8918b990ea7ded1088cb7a88693ca480728d2130b78de63" exitCode=0 Mar 16 00:22:56 crc kubenswrapper[4751]: I0316 00:22:56.271941 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"32d5f19b-e735-49bd-8212-3367fd97cc49","Type":"ContainerDied","Data":"7a1b7f8df8e635ade8918b990ea7ded1088cb7a88693ca480728d2130b78de63"} Mar 16 00:22:57 crc kubenswrapper[4751]: I0316 00:22:57.067795 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:57 crc kubenswrapper[4751]: I0316 00:22:57.073580 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/24d89dab-af8f-4823-b6a4-37e9fa41a466-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp\" (UID: \"24d89dab-af8f-4823-b6a4-37e9fa41a466\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:57 crc kubenswrapper[4751]: I0316 00:22:57.206314 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" Mar 16 00:22:57 crc kubenswrapper[4751]: I0316 00:22:57.812458 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp"] Mar 16 00:22:57 crc kubenswrapper[4751]: I0316 00:22:57.909397 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk"] Mar 16 00:22:58 crc kubenswrapper[4751]: I0316 00:22:58.037984 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq"] Mar 16 00:22:58 crc kubenswrapper[4751]: W0316 00:22:58.206465 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93427807_2b70_4cbe_9e6e_6fc6486a0020.slice/crio-d52d1290e4ad3400460142e7bc453bd8819368c18dd527beea748c65f180b5b7 WatchSource:0}: Error finding container d52d1290e4ad3400460142e7bc453bd8819368c18dd527beea748c65f180b5b7: Status 404 returned error can't find the container with id d52d1290e4ad3400460142e7bc453bd8819368c18dd527beea748c65f180b5b7 Mar 16 00:22:58 crc kubenswrapper[4751]: I0316 00:22:58.287444 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" event={"ID":"93427807-2b70-4cbe-9e6e-6fc6486a0020","Type":"ContainerStarted","Data":"d52d1290e4ad3400460142e7bc453bd8819368c18dd527beea748c65f180b5b7"} Mar 16 00:22:58 crc kubenswrapper[4751]: I0316 00:22:58.288324 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" event={"ID":"ccf828f2-8f37-48da-8f95-b0e0ce744f34","Type":"ContainerStarted","Data":"4e07fa5d5e6bda11ccafc1eecce98d8d1a24177597fbf8a7a4289e4d3a2eead4"} Mar 16 00:22:58 crc kubenswrapper[4751]: I0316 00:22:58.292629 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"aaf54795-c77b-42bf-a088-5104d905ab2d","Type":"ContainerStarted","Data":"2b05fe48ae1bdbf51135fa86f9e7f3ae5a5a38fc00e63c74500b2a03cef2e2d2"} Mar 16 00:22:58 crc kubenswrapper[4751]: I0316 00:22:58.297040 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" event={"ID":"24d89dab-af8f-4823-b6a4-37e9fa41a466","Type":"ContainerStarted","Data":"4cc3b3a09cfd3be666c2cdaea252bc142caf0b2ce58be037f852869ea6d82de7"} Mar 16 00:22:58 crc kubenswrapper[4751]: I0316 00:22:58.320195 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=4.944794275 podStartE2EDuration="37.320179858s" podCreationTimestamp="2026-03-16 00:22:21 +0000 UTC" firstStartedPulling="2026-03-16 00:22:25.257520312 +0000 UTC m=+973.526852592" lastFinishedPulling="2026-03-16 00:22:57.632905885 +0000 UTC m=+1005.902238175" observedRunningTime="2026-03-16 00:22:58.313501711 +0000 UTC m=+1006.582834001" watchObservedRunningTime="2026-03-16 00:22:58.320179858 +0000 UTC m=+1006.589512149" Mar 16 00:22:59 crc kubenswrapper[4751]: I0316 00:22:59.308373 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" event={"ID":"24d89dab-af8f-4823-b6a4-37e9fa41a466","Type":"ContainerStarted","Data":"01ed45e4e014e57111d877e1e0e8f43d2f176e48ffae617453d2b56b2c50c603"} Mar 16 00:22:59 crc kubenswrapper[4751]: I0316 00:22:59.769666 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/prometheus-default-0" Mar 16 00:23:00 crc kubenswrapper[4751]: I0316 00:23:00.316306 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"32d5f19b-e735-49bd-8212-3367fd97cc49","Type":"ContainerStarted","Data":"17c6d07419f5c213173af11939fe8ef3298977f26ae6ec4200520724c0991aa7"} Mar 16 00:23:00 crc kubenswrapper[4751]: I0316 00:23:00.318434 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" event={"ID":"93427807-2b70-4cbe-9e6e-6fc6486a0020","Type":"ContainerStarted","Data":"fb66e273e272df028986fc2e0c1df61f3abf018ad6c46691baa29b6852db3f06"} Mar 16 00:23:00 crc kubenswrapper[4751]: I0316 00:23:00.319926 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" event={"ID":"ccf828f2-8f37-48da-8f95-b0e0ce744f34","Type":"ContainerStarted","Data":"c34433a9b5220f93b294145c2b9d5dd47a62c4936477873b4349f05956d0a0b6"} Mar 16 00:23:01 crc kubenswrapper[4751]: I0316 00:23:01.329714 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"32d5f19b-e735-49bd-8212-3367fd97cc49","Type":"ContainerStarted","Data":"1c4668a7c326f2412df34529fe3d1b70dbeca76cb5ea8557764324090011e425"} Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.339414 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"32d5f19b-e735-49bd-8212-3367fd97cc49","Type":"ContainerStarted","Data":"04b787f89a5ba5378863185e81d0369959d7f471962f1dc05d710a797e49bf0c"} Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.388745 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=21.758048739 podStartE2EDuration="27.388720137s" podCreationTimestamp="2026-03-16 00:22:35 +0000 UTC" firstStartedPulling="2026-03-16 00:22:56.274904821 +0000 UTC m=+1004.544237111" lastFinishedPulling="2026-03-16 00:23:01.905576219 +0000 UTC m=+1010.174908509" observedRunningTime="2026-03-16 00:23:02.366219064 +0000 UTC m=+1010.635551354" watchObservedRunningTime="2026-03-16 00:23:02.388720137 +0000 UTC m=+1010.658052427" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.554836 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4"] Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.555835 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.559078 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-cert" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.559342 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-event-sg-core-configmap" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.561253 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4"] Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.650048 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/87bbd13c-ea3b-4664-bafe-92ac339d62ef-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.650134 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/87bbd13c-ea3b-4664-bafe-92ac339d62ef-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.650160 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p56b\" (UniqueName: \"kubernetes.io/projected/87bbd13c-ea3b-4664-bafe-92ac339d62ef-kube-api-access-6p56b\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.650187 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/87bbd13c-ea3b-4664-bafe-92ac339d62ef-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.751599 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/87bbd13c-ea3b-4664-bafe-92ac339d62ef-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.751638 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p56b\" (UniqueName: \"kubernetes.io/projected/87bbd13c-ea3b-4664-bafe-92ac339d62ef-kube-api-access-6p56b\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.751673 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/87bbd13c-ea3b-4664-bafe-92ac339d62ef-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.751737 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/87bbd13c-ea3b-4664-bafe-92ac339d62ef-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.752547 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/87bbd13c-ea3b-4664-bafe-92ac339d62ef-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.760265 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/87bbd13c-ea3b-4664-bafe-92ac339d62ef-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.762508 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/87bbd13c-ea3b-4664-bafe-92ac339d62ef-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.777576 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p56b\" (UniqueName: \"kubernetes.io/projected/87bbd13c-ea3b-4664-bafe-92ac339d62ef-kube-api-access-6p56b\") pod \"default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4\" (UID: \"87bbd13c-ea3b-4664-bafe-92ac339d62ef\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:02 crc kubenswrapper[4751]: I0316 00:23:02.872849 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.169172 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6"] Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.173861 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.176430 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-event-sg-core-configmap" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.176687 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6"] Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.363131 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/3229d246-5e60-4ae7-8d1a-0fcf116d4038-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.363326 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b7jz\" (UniqueName: \"kubernetes.io/projected/3229d246-5e60-4ae7-8d1a-0fcf116d4038-kube-api-access-4b7jz\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.363398 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/3229d246-5e60-4ae7-8d1a-0fcf116d4038-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.363466 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/3229d246-5e60-4ae7-8d1a-0fcf116d4038-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.464248 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/3229d246-5e60-4ae7-8d1a-0fcf116d4038-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.464329 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b7jz\" (UniqueName: \"kubernetes.io/projected/3229d246-5e60-4ae7-8d1a-0fcf116d4038-kube-api-access-4b7jz\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.464362 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/3229d246-5e60-4ae7-8d1a-0fcf116d4038-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.464390 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/3229d246-5e60-4ae7-8d1a-0fcf116d4038-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.465977 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/3229d246-5e60-4ae7-8d1a-0fcf116d4038-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.466022 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/3229d246-5e60-4ae7-8d1a-0fcf116d4038-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.469455 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/3229d246-5e60-4ae7-8d1a-0fcf116d4038-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.479660 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b7jz\" (UniqueName: \"kubernetes.io/projected/3229d246-5e60-4ae7-8d1a-0fcf116d4038-kube-api-access-4b7jz\") pod \"default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6\" (UID: \"3229d246-5e60-4ae7-8d1a-0fcf116d4038\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:03 crc kubenswrapper[4751]: I0316 00:23:03.491317 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" Mar 16 00:23:07 crc kubenswrapper[4751]: I0316 00:23:07.377737 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" event={"ID":"ccf828f2-8f37-48da-8f95-b0e0ce744f34","Type":"ContainerStarted","Data":"5715b819f6de6f6816cee4fc1842c93f384e424b1446002aec9a980669b3dc50"} Mar 16 00:23:07 crc kubenswrapper[4751]: I0316 00:23:07.380155 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" event={"ID":"24d89dab-af8f-4823-b6a4-37e9fa41a466","Type":"ContainerStarted","Data":"9c6581aeca49c0af9064a2e12bfa468cafc754a79c57e7e242f520ed04832946"} Mar 16 00:23:07 crc kubenswrapper[4751]: I0316 00:23:07.381671 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" event={"ID":"93427807-2b70-4cbe-9e6e-6fc6486a0020","Type":"ContainerStarted","Data":"262432c06a8869633a3eea116351523ccae89967e6279e3235c7c0f1f29d7590"} Mar 16 00:23:07 crc kubenswrapper[4751]: I0316 00:23:07.448043 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4"] Mar 16 00:23:07 crc kubenswrapper[4751]: W0316 00:23:07.448618 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87bbd13c_ea3b_4664_bafe_92ac339d62ef.slice/crio-49c7953313696e228f7136321fa23d8d5678cc6699bade20a5f6824a742f14e6 WatchSource:0}: Error finding container 49c7953313696e228f7136321fa23d8d5678cc6699bade20a5f6824a742f14e6: Status 404 returned error can't find the container with id 49c7953313696e228f7136321fa23d8d5678cc6699bade20a5f6824a742f14e6 Mar 16 00:23:07 crc kubenswrapper[4751]: I0316 00:23:07.584769 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6"] Mar 16 00:23:07 crc kubenswrapper[4751]: W0316 00:23:07.600400 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3229d246_5e60_4ae7_8d1a_0fcf116d4038.slice/crio-2a189149079b6f259f9041e17d0095914df6d9e287d33170c2aed416d1bb720c WatchSource:0}: Error finding container 2a189149079b6f259f9041e17d0095914df6d9e287d33170c2aed416d1bb720c: Status 404 returned error can't find the container with id 2a189149079b6f259f9041e17d0095914df6d9e287d33170c2aed416d1bb720c Mar 16 00:23:08 crc kubenswrapper[4751]: I0316 00:23:08.393577 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" event={"ID":"3229d246-5e60-4ae7-8d1a-0fcf116d4038","Type":"ContainerStarted","Data":"9653ebf65b6300069c05765b3bff97406e722f11b6548a243904ba0ad2484535"} Mar 16 00:23:08 crc kubenswrapper[4751]: I0316 00:23:08.394176 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" event={"ID":"3229d246-5e60-4ae7-8d1a-0fcf116d4038","Type":"ContainerStarted","Data":"2a189149079b6f259f9041e17d0095914df6d9e287d33170c2aed416d1bb720c"} Mar 16 00:23:08 crc kubenswrapper[4751]: I0316 00:23:08.405361 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" event={"ID":"87bbd13c-ea3b-4664-bafe-92ac339d62ef","Type":"ContainerStarted","Data":"06f96c18154ce93c7bd2ae51d14819984445e2f7b19d906c05a7b97ac365bc30"} Mar 16 00:23:08 crc kubenswrapper[4751]: I0316 00:23:08.405460 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" event={"ID":"87bbd13c-ea3b-4664-bafe-92ac339d62ef","Type":"ContainerStarted","Data":"49c7953313696e228f7136321fa23d8d5678cc6699bade20a5f6824a742f14e6"} Mar 16 00:23:09 crc kubenswrapper[4751]: I0316 00:23:09.770258 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Mar 16 00:23:09 crc kubenswrapper[4751]: I0316 00:23:09.822486 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Mar 16 00:23:10 crc kubenswrapper[4751]: I0316 00:23:10.473600 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.169381 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xtrt6"] Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.170035 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" podUID="58867d99-7102-43c5-9e18-eb59e2c702b6" containerName="default-interconnect" containerID="cri-o://fbc2394f9179ee1cc977dd5f4fa92be4efab1cbb916ee226b3a509e55797b5f1" gracePeriod=30 Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.452694 4751 generic.go:334] "Generic (PLEG): container finished" podID="24d89dab-af8f-4823-b6a4-37e9fa41a466" containerID="9c6581aeca49c0af9064a2e12bfa468cafc754a79c57e7e242f520ed04832946" exitCode=0 Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.452750 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" event={"ID":"24d89dab-af8f-4823-b6a4-37e9fa41a466","Type":"ContainerDied","Data":"9c6581aeca49c0af9064a2e12bfa468cafc754a79c57e7e242f520ed04832946"} Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.454720 4751 generic.go:334] "Generic (PLEG): container finished" podID="93427807-2b70-4cbe-9e6e-6fc6486a0020" containerID="262432c06a8869633a3eea116351523ccae89967e6279e3235c7c0f1f29d7590" exitCode=0 Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.456624 4751 generic.go:334] "Generic (PLEG): container finished" podID="87bbd13c-ea3b-4664-bafe-92ac339d62ef" containerID="06f96c18154ce93c7bd2ae51d14819984445e2f7b19d906c05a7b97ac365bc30" exitCode=0 Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.458780 4751 generic.go:334] "Generic (PLEG): container finished" podID="58867d99-7102-43c5-9e18-eb59e2c702b6" containerID="fbc2394f9179ee1cc977dd5f4fa92be4efab1cbb916ee226b3a509e55797b5f1" exitCode=0 Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.460215 4751 generic.go:334] "Generic (PLEG): container finished" podID="ccf828f2-8f37-48da-8f95-b0e0ce744f34" containerID="5715b819f6de6f6816cee4fc1842c93f384e424b1446002aec9a980669b3dc50" exitCode=0 Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.462050 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" event={"ID":"93427807-2b70-4cbe-9e6e-6fc6486a0020","Type":"ContainerDied","Data":"262432c06a8869633a3eea116351523ccae89967e6279e3235c7c0f1f29d7590"} Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.462079 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" event={"ID":"87bbd13c-ea3b-4664-bafe-92ac339d62ef","Type":"ContainerDied","Data":"06f96c18154ce93c7bd2ae51d14819984445e2f7b19d906c05a7b97ac365bc30"} Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.462091 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" event={"ID":"58867d99-7102-43c5-9e18-eb59e2c702b6","Type":"ContainerDied","Data":"fbc2394f9179ee1cc977dd5f4fa92be4efab1cbb916ee226b3a509e55797b5f1"} Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.462104 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" event={"ID":"ccf828f2-8f37-48da-8f95-b0e0ce744f34","Type":"ContainerDied","Data":"5715b819f6de6f6816cee4fc1842c93f384e424b1446002aec9a980669b3dc50"} Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.462213 4751 generic.go:334] "Generic (PLEG): container finished" podID="3229d246-5e60-4ae7-8d1a-0fcf116d4038" containerID="9653ebf65b6300069c05765b3bff97406e722f11b6548a243904ba0ad2484535" exitCode=0 Mar 16 00:23:14 crc kubenswrapper[4751]: I0316 00:23:14.462232 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" event={"ID":"3229d246-5e60-4ae7-8d1a-0fcf116d4038","Type":"ContainerDied","Data":"9653ebf65b6300069c05765b3bff97406e722f11b6548a243904ba0ad2484535"} Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.834961 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.878658 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-2trgn"] Mar 16 00:23:15 crc kubenswrapper[4751]: E0316 00:23:15.879035 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58867d99-7102-43c5-9e18-eb59e2c702b6" containerName="default-interconnect" Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.879124 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="58867d99-7102-43c5-9e18-eb59e2c702b6" containerName="default-interconnect" Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.879365 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="58867d99-7102-43c5-9e18-eb59e2c702b6" containerName="default-interconnect" Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.879950 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.898343 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-2trgn"] Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.998978 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-ca\") pod \"58867d99-7102-43c5-9e18-eb59e2c702b6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.999043 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-credentials\") pod \"58867d99-7102-43c5-9e18-eb59e2c702b6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.999068 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-credentials\") pod \"58867d99-7102-43c5-9e18-eb59e2c702b6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.999151 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-ca\") pod \"58867d99-7102-43c5-9e18-eb59e2c702b6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.999173 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-users\") pod \"58867d99-7102-43c5-9e18-eb59e2c702b6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.999724 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-config\") pod \"58867d99-7102-43c5-9e18-eb59e2c702b6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " Mar 16 00:23:15 crc kubenswrapper[4751]: I0316 00:23:15.999767 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkkmz\" (UniqueName: \"kubernetes.io/projected/58867d99-7102-43c5-9e18-eb59e2c702b6-kube-api-access-kkkmz\") pod \"58867d99-7102-43c5-9e18-eb59e2c702b6\" (UID: \"58867d99-7102-43c5-9e18-eb59e2c702b6\") " Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:15.999977 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.000022 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.000051 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-sasl-users\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.000090 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.000159 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.000207 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/4054f75c-cc02-418d-a301-e3a7c92ef630-sasl-config\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.000234 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdx54\" (UniqueName: \"kubernetes.io/projected/4054f75c-cc02-418d-a301-e3a7c92ef630-kube-api-access-kdx54\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.000913 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "58867d99-7102-43c5-9e18-eb59e2c702b6" (UID: "58867d99-7102-43c5-9e18-eb59e2c702b6"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.003762 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "58867d99-7102-43c5-9e18-eb59e2c702b6" (UID: "58867d99-7102-43c5-9e18-eb59e2c702b6"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.004321 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58867d99-7102-43c5-9e18-eb59e2c702b6-kube-api-access-kkkmz" (OuterVolumeSpecName: "kube-api-access-kkkmz") pod "58867d99-7102-43c5-9e18-eb59e2c702b6" (UID: "58867d99-7102-43c5-9e18-eb59e2c702b6"). InnerVolumeSpecName "kube-api-access-kkkmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.004398 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "58867d99-7102-43c5-9e18-eb59e2c702b6" (UID: "58867d99-7102-43c5-9e18-eb59e2c702b6"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.005051 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "58867d99-7102-43c5-9e18-eb59e2c702b6" (UID: "58867d99-7102-43c5-9e18-eb59e2c702b6"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.005228 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "58867d99-7102-43c5-9e18-eb59e2c702b6" (UID: "58867d99-7102-43c5-9e18-eb59e2c702b6"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.007288 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "58867d99-7102-43c5-9e18-eb59e2c702b6" (UID: "58867d99-7102-43c5-9e18-eb59e2c702b6"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.100888 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/4054f75c-cc02-418d-a301-e3a7c92ef630-sasl-config\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.100933 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdx54\" (UniqueName: \"kubernetes.io/projected/4054f75c-cc02-418d-a301-e3a7c92ef630-kube-api-access-kdx54\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.100983 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101013 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101031 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-sasl-users\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101056 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101084 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101152 4751 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101164 4751 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101176 4751 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101185 4751 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101194 4751 reconciler_common.go:293] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-users\") on node \"crc\" DevicePath \"\"" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101204 4751 reconciler_common.go:293] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/58867d99-7102-43c5-9e18-eb59e2c702b6-sasl-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.101212 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkkmz\" (UniqueName: \"kubernetes.io/projected/58867d99-7102-43c5-9e18-eb59e2c702b6-kube-api-access-kkkmz\") on node \"crc\" DevicePath \"\"" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.103975 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/4054f75c-cc02-418d-a301-e3a7c92ef630-sasl-config\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.107422 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.108090 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.108130 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.113993 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.114345 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/4054f75c-cc02-418d-a301-e3a7c92ef630-sasl-users\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.124849 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdx54\" (UniqueName: \"kubernetes.io/projected/4054f75c-cc02-418d-a301-e3a7c92ef630-kube-api-access-kdx54\") pod \"default-interconnect-68864d46cb-2trgn\" (UID: \"4054f75c-cc02-418d-a301-e3a7c92ef630\") " pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.217526 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-2trgn" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.480688 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.481187 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xtrt6" event={"ID":"58867d99-7102-43c5-9e18-eb59e2c702b6","Type":"ContainerDied","Data":"2bcc8dec9dada6e5bc5b7049070c865bf6fc0782debf5eb75c3e652bb1448dac"} Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.481230 4751 scope.go:117] "RemoveContainer" containerID="fbc2394f9179ee1cc977dd5f4fa92be4efab1cbb916ee226b3a509e55797b5f1" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.487227 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" event={"ID":"ccf828f2-8f37-48da-8f95-b0e0ce744f34","Type":"ContainerStarted","Data":"14a034fc39031f8d59a4ba8a3af8ec96a5e58a03ca06d8d4af10887afb8744fe"} Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.488828 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" event={"ID":"3229d246-5e60-4ae7-8d1a-0fcf116d4038","Type":"ContainerStarted","Data":"07e6306de0f92b711628abbc11e1a9774934c228c159728303b862d9a1dfc3fe"} Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.489652 4751 scope.go:117] "RemoveContainer" containerID="9653ebf65b6300069c05765b3bff97406e722f11b6548a243904ba0ad2484535" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.504714 4751 scope.go:117] "RemoveContainer" containerID="5715b819f6de6f6816cee4fc1842c93f384e424b1446002aec9a980669b3dc50" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.514088 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" event={"ID":"24d89dab-af8f-4823-b6a4-37e9fa41a466","Type":"ContainerStarted","Data":"da375f50416e29e8d5fa8757b9ab398a14a993f316782c37e3f6a7ac8eaab726"} Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.517964 4751 scope.go:117] "RemoveContainer" containerID="9c6581aeca49c0af9064a2e12bfa468cafc754a79c57e7e242f520ed04832946" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.520209 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" event={"ID":"93427807-2b70-4cbe-9e6e-6fc6486a0020","Type":"ContainerStarted","Data":"fe124393adb9a1d67d8cbe933dad6c6ee464df28da0cea22ea85c57c041e10ef"} Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.521243 4751 scope.go:117] "RemoveContainer" containerID="262432c06a8869633a3eea116351523ccae89967e6279e3235c7c0f1f29d7590" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.524369 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" event={"ID":"87bbd13c-ea3b-4664-bafe-92ac339d62ef","Type":"ContainerStarted","Data":"ab9b1b16bff2209cbcc90d79917f2737ea655a0fe2d99492bd1fdff04c3cfac8"} Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.527701 4751 scope.go:117] "RemoveContainer" containerID="06f96c18154ce93c7bd2ae51d14819984445e2f7b19d906c05a7b97ac365bc30" Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.536146 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xtrt6"] Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.550336 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xtrt6"] Mar 16 00:23:16 crc kubenswrapper[4751]: I0316 00:23:16.633985 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-2trgn"] Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.540379 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" event={"ID":"93427807-2b70-4cbe-9e6e-6fc6486a0020","Type":"ContainerStarted","Data":"73382b4d68e09937b679d0d640ec13f0225a86850a1abf61fce61f8528294b4e"} Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.548080 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" event={"ID":"87bbd13c-ea3b-4664-bafe-92ac339d62ef","Type":"ContainerStarted","Data":"3b04832e7795cbb8ad926bac73ce87de233de3a8e525b1a119d18ee4c68ebde8"} Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.552739 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-2trgn" event={"ID":"4054f75c-cc02-418d-a301-e3a7c92ef630","Type":"ContainerStarted","Data":"d212e31d1cc3ba6c6bc08a49da5eb8fa876990d751262608f0781194778960e5"} Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.552973 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-2trgn" event={"ID":"4054f75c-cc02-418d-a301-e3a7c92ef630","Type":"ContainerStarted","Data":"d3e9c7ec26dfa5ec532fd476362a948b1dd467ed5c6632dd038f215e8a0cacb1"} Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.560691 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" event={"ID":"ccf828f2-8f37-48da-8f95-b0e0ce744f34","Type":"ContainerStarted","Data":"bfd7b74e7f8cb0313d6cdbc197a083cc75f2c4a2f1eaa292c92dce0968d3205f"} Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.571303 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" podStartSLOduration=9.80698854 podStartE2EDuration="28.571273041s" podCreationTimestamp="2026-03-16 00:22:49 +0000 UTC" firstStartedPulling="2026-03-16 00:22:58.210371573 +0000 UTC m=+1006.479703863" lastFinishedPulling="2026-03-16 00:23:16.974656074 +0000 UTC m=+1025.243988364" observedRunningTime="2026-03-16 00:23:17.561902598 +0000 UTC m=+1025.831234928" watchObservedRunningTime="2026-03-16 00:23:17.571273041 +0000 UTC m=+1025.840605381" Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.572851 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" event={"ID":"3229d246-5e60-4ae7-8d1a-0fcf116d4038","Type":"ContainerStarted","Data":"caf2db675fa55e2ad6e938a5a36435079ade14e6fe942460c7fdafdc8f6cb3c5"} Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.596842 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" event={"ID":"24d89dab-af8f-4823-b6a4-37e9fa41a466","Type":"ContainerStarted","Data":"fc969b536c76c025ccca9dfa9fbdec2a9b5478db1e634c4a3ce7f89d2e997033"} Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.683078 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-2trgn" podStartSLOduration=3.683057373 podStartE2EDuration="3.683057373s" podCreationTimestamp="2026-03-16 00:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-16 00:23:17.65521092 +0000 UTC m=+1025.924543300" watchObservedRunningTime="2026-03-16 00:23:17.683057373 +0000 UTC m=+1025.952389673" Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.704204 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" podStartSLOduration=8.006942648 podStartE2EDuration="26.704179996s" podCreationTimestamp="2026-03-16 00:22:51 +0000 UTC" firstStartedPulling="2026-03-16 00:22:58.215405534 +0000 UTC m=+1006.484737824" lastFinishedPulling="2026-03-16 00:23:16.912642882 +0000 UTC m=+1025.181975172" observedRunningTime="2026-03-16 00:23:17.702673254 +0000 UTC m=+1025.972005554" watchObservedRunningTime="2026-03-16 00:23:17.704179996 +0000 UTC m=+1025.973512296" Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.712057 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" podStartSLOduration=6.116812383 podStartE2EDuration="15.712037377s" podCreationTimestamp="2026-03-16 00:23:02 +0000 UTC" firstStartedPulling="2026-03-16 00:23:07.452219846 +0000 UTC m=+1015.721552136" lastFinishedPulling="2026-03-16 00:23:17.04744484 +0000 UTC m=+1025.316777130" observedRunningTime="2026-03-16 00:23:17.681491229 +0000 UTC m=+1025.950823539" watchObservedRunningTime="2026-03-16 00:23:17.712037377 +0000 UTC m=+1025.981369677" Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.733185 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" podStartSLOduration=3.603536092 podStartE2EDuration="22.733166761s" podCreationTimestamp="2026-03-16 00:22:55 +0000 UTC" firstStartedPulling="2026-03-16 00:22:57.845252932 +0000 UTC m=+1006.114585222" lastFinishedPulling="2026-03-16 00:23:16.974883601 +0000 UTC m=+1025.244215891" observedRunningTime="2026-03-16 00:23:17.723935191 +0000 UTC m=+1025.993267501" watchObservedRunningTime="2026-03-16 00:23:17.733166761 +0000 UTC m=+1026.002499051" Mar 16 00:23:17 crc kubenswrapper[4751]: I0316 00:23:17.755814 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" podStartSLOduration=5.391302698 podStartE2EDuration="14.755799147s" podCreationTimestamp="2026-03-16 00:23:03 +0000 UTC" firstStartedPulling="2026-03-16 00:23:07.6082046 +0000 UTC m=+1015.877536890" lastFinishedPulling="2026-03-16 00:23:16.972701049 +0000 UTC m=+1025.242033339" observedRunningTime="2026-03-16 00:23:17.752727201 +0000 UTC m=+1026.022059501" watchObservedRunningTime="2026-03-16 00:23:17.755799147 +0000 UTC m=+1026.025131427" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.463939 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58867d99-7102-43c5-9e18-eb59e2c702b6" path="/var/lib/kubelet/pods/58867d99-7102-43c5-9e18-eb59e2c702b6/volumes" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.607335 4751 generic.go:334] "Generic (PLEG): container finished" podID="93427807-2b70-4cbe-9e6e-6fc6486a0020" containerID="73382b4d68e09937b679d0d640ec13f0225a86850a1abf61fce61f8528294b4e" exitCode=0 Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.607408 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" event={"ID":"93427807-2b70-4cbe-9e6e-6fc6486a0020","Type":"ContainerDied","Data":"73382b4d68e09937b679d0d640ec13f0225a86850a1abf61fce61f8528294b4e"} Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.607454 4751 scope.go:117] "RemoveContainer" containerID="262432c06a8869633a3eea116351523ccae89967e6279e3235c7c0f1f29d7590" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.608009 4751 scope.go:117] "RemoveContainer" containerID="73382b4d68e09937b679d0d640ec13f0225a86850a1abf61fce61f8528294b4e" Mar 16 00:23:18 crc kubenswrapper[4751]: E0316 00:23:18.608240 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq_service-telemetry(93427807-2b70-4cbe-9e6e-6fc6486a0020)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" podUID="93427807-2b70-4cbe-9e6e-6fc6486a0020" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.615668 4751 generic.go:334] "Generic (PLEG): container finished" podID="87bbd13c-ea3b-4664-bafe-92ac339d62ef" containerID="3b04832e7795cbb8ad926bac73ce87de233de3a8e525b1a119d18ee4c68ebde8" exitCode=0 Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.615731 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" event={"ID":"87bbd13c-ea3b-4664-bafe-92ac339d62ef","Type":"ContainerDied","Data":"3b04832e7795cbb8ad926bac73ce87de233de3a8e525b1a119d18ee4c68ebde8"} Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.616166 4751 scope.go:117] "RemoveContainer" containerID="3b04832e7795cbb8ad926bac73ce87de233de3a8e525b1a119d18ee4c68ebde8" Mar 16 00:23:18 crc kubenswrapper[4751]: E0316 00:23:18.616372 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4_service-telemetry(87bbd13c-ea3b-4664-bafe-92ac339d62ef)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" podUID="87bbd13c-ea3b-4664-bafe-92ac339d62ef" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.619164 4751 generic.go:334] "Generic (PLEG): container finished" podID="ccf828f2-8f37-48da-8f95-b0e0ce744f34" containerID="bfd7b74e7f8cb0313d6cdbc197a083cc75f2c4a2f1eaa292c92dce0968d3205f" exitCode=0 Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.619227 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" event={"ID":"ccf828f2-8f37-48da-8f95-b0e0ce744f34","Type":"ContainerDied","Data":"bfd7b74e7f8cb0313d6cdbc197a083cc75f2c4a2f1eaa292c92dce0968d3205f"} Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.619789 4751 scope.go:117] "RemoveContainer" containerID="bfd7b74e7f8cb0313d6cdbc197a083cc75f2c4a2f1eaa292c92dce0968d3205f" Mar 16 00:23:18 crc kubenswrapper[4751]: E0316 00:23:18.620033 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk_service-telemetry(ccf828f2-8f37-48da-8f95-b0e0ce744f34)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" podUID="ccf828f2-8f37-48da-8f95-b0e0ce744f34" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.621540 4751 generic.go:334] "Generic (PLEG): container finished" podID="3229d246-5e60-4ae7-8d1a-0fcf116d4038" containerID="caf2db675fa55e2ad6e938a5a36435079ade14e6fe942460c7fdafdc8f6cb3c5" exitCode=0 Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.621588 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" event={"ID":"3229d246-5e60-4ae7-8d1a-0fcf116d4038","Type":"ContainerDied","Data":"caf2db675fa55e2ad6e938a5a36435079ade14e6fe942460c7fdafdc8f6cb3c5"} Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.621904 4751 scope.go:117] "RemoveContainer" containerID="caf2db675fa55e2ad6e938a5a36435079ade14e6fe942460c7fdafdc8f6cb3c5" Mar 16 00:23:18 crc kubenswrapper[4751]: E0316 00:23:18.622148 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6_service-telemetry(3229d246-5e60-4ae7-8d1a-0fcf116d4038)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" podUID="3229d246-5e60-4ae7-8d1a-0fcf116d4038" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.625215 4751 generic.go:334] "Generic (PLEG): container finished" podID="24d89dab-af8f-4823-b6a4-37e9fa41a466" containerID="fc969b536c76c025ccca9dfa9fbdec2a9b5478db1e634c4a3ce7f89d2e997033" exitCode=0 Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.626033 4751 scope.go:117] "RemoveContainer" containerID="fc969b536c76c025ccca9dfa9fbdec2a9b5478db1e634c4a3ce7f89d2e997033" Mar 16 00:23:18 crc kubenswrapper[4751]: E0316 00:23:18.626230 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp_service-telemetry(24d89dab-af8f-4823-b6a4-37e9fa41a466)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" podUID="24d89dab-af8f-4823-b6a4-37e9fa41a466" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.626263 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" event={"ID":"24d89dab-af8f-4823-b6a4-37e9fa41a466","Type":"ContainerDied","Data":"fc969b536c76c025ccca9dfa9fbdec2a9b5478db1e634c4a3ce7f89d2e997033"} Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.690095 4751 scope.go:117] "RemoveContainer" containerID="06f96c18154ce93c7bd2ae51d14819984445e2f7b19d906c05a7b97ac365bc30" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.740460 4751 scope.go:117] "RemoveContainer" containerID="5715b819f6de6f6816cee4fc1842c93f384e424b1446002aec9a980669b3dc50" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.787355 4751 scope.go:117] "RemoveContainer" containerID="9653ebf65b6300069c05765b3bff97406e722f11b6548a243904ba0ad2484535" Mar 16 00:23:18 crc kubenswrapper[4751]: I0316 00:23:18.825306 4751 scope.go:117] "RemoveContainer" containerID="9c6581aeca49c0af9064a2e12bfa468cafc754a79c57e7e242f520ed04832946" Mar 16 00:23:19 crc kubenswrapper[4751]: I0316 00:23:19.634958 4751 scope.go:117] "RemoveContainer" containerID="bfd7b74e7f8cb0313d6cdbc197a083cc75f2c4a2f1eaa292c92dce0968d3205f" Mar 16 00:23:19 crc kubenswrapper[4751]: E0316 00:23:19.635378 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk_service-telemetry(ccf828f2-8f37-48da-8f95-b0e0ce744f34)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" podUID="ccf828f2-8f37-48da-8f95-b0e0ce744f34" Mar 16 00:23:19 crc kubenswrapper[4751]: I0316 00:23:19.637023 4751 scope.go:117] "RemoveContainer" containerID="caf2db675fa55e2ad6e938a5a36435079ade14e6fe942460c7fdafdc8f6cb3c5" Mar 16 00:23:19 crc kubenswrapper[4751]: E0316 00:23:19.637383 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6_service-telemetry(3229d246-5e60-4ae7-8d1a-0fcf116d4038)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" podUID="3229d246-5e60-4ae7-8d1a-0fcf116d4038" Mar 16 00:23:19 crc kubenswrapper[4751]: I0316 00:23:19.639393 4751 scope.go:117] "RemoveContainer" containerID="fc969b536c76c025ccca9dfa9fbdec2a9b5478db1e634c4a3ce7f89d2e997033" Mar 16 00:23:19 crc kubenswrapper[4751]: E0316 00:23:19.639707 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp_service-telemetry(24d89dab-af8f-4823-b6a4-37e9fa41a466)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" podUID="24d89dab-af8f-4823-b6a4-37e9fa41a466" Mar 16 00:23:19 crc kubenswrapper[4751]: I0316 00:23:19.641382 4751 scope.go:117] "RemoveContainer" containerID="73382b4d68e09937b679d0d640ec13f0225a86850a1abf61fce61f8528294b4e" Mar 16 00:23:19 crc kubenswrapper[4751]: E0316 00:23:19.641766 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq_service-telemetry(93427807-2b70-4cbe-9e6e-6fc6486a0020)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" podUID="93427807-2b70-4cbe-9e6e-6fc6486a0020" Mar 16 00:23:19 crc kubenswrapper[4751]: I0316 00:23:19.643280 4751 scope.go:117] "RemoveContainer" containerID="3b04832e7795cbb8ad926bac73ce87de233de3a8e525b1a119d18ee4c68ebde8" Mar 16 00:23:19 crc kubenswrapper[4751]: E0316 00:23:19.643500 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4_service-telemetry(87bbd13c-ea3b-4664-bafe-92ac339d62ef)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" podUID="87bbd13c-ea3b-4664-bafe-92ac339d62ef" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.429354 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.430797 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.461091 4751 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-selfsigned" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.461212 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"qdr-test-config" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.467179 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.586388 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzq2r\" (UniqueName: \"kubernetes.io/projected/af16bec7-f32d-4931-b196-d02ecd52cd3e-kube-api-access-qzq2r\") pod \"qdr-test\" (UID: \"af16bec7-f32d-4931-b196-d02ecd52cd3e\") " pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.586478 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/af16bec7-f32d-4931-b196-d02ecd52cd3e-qdr-test-config\") pod \"qdr-test\" (UID: \"af16bec7-f32d-4931-b196-d02ecd52cd3e\") " pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.586534 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/af16bec7-f32d-4931-b196-d02ecd52cd3e-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"af16bec7-f32d-4931-b196-d02ecd52cd3e\") " pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.687801 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzq2r\" (UniqueName: \"kubernetes.io/projected/af16bec7-f32d-4931-b196-d02ecd52cd3e-kube-api-access-qzq2r\") pod \"qdr-test\" (UID: \"af16bec7-f32d-4931-b196-d02ecd52cd3e\") " pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.688181 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/af16bec7-f32d-4931-b196-d02ecd52cd3e-qdr-test-config\") pod \"qdr-test\" (UID: \"af16bec7-f32d-4931-b196-d02ecd52cd3e\") " pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.688227 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/af16bec7-f32d-4931-b196-d02ecd52cd3e-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"af16bec7-f32d-4931-b196-d02ecd52cd3e\") " pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.689648 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/af16bec7-f32d-4931-b196-d02ecd52cd3e-qdr-test-config\") pod \"qdr-test\" (UID: \"af16bec7-f32d-4931-b196-d02ecd52cd3e\") " pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.694413 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/af16bec7-f32d-4931-b196-d02ecd52cd3e-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"af16bec7-f32d-4931-b196-d02ecd52cd3e\") " pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.727662 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzq2r\" (UniqueName: \"kubernetes.io/projected/af16bec7-f32d-4931-b196-d02ecd52cd3e-kube-api-access-qzq2r\") pod \"qdr-test\" (UID: \"af16bec7-f32d-4931-b196-d02ecd52cd3e\") " pod="service-telemetry/qdr-test" Mar 16 00:23:27 crc kubenswrapper[4751]: I0316 00:23:27.786010 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 16 00:23:28 crc kubenswrapper[4751]: I0316 00:23:28.223881 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 16 00:23:28 crc kubenswrapper[4751]: W0316 00:23:28.226254 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf16bec7_f32d_4931_b196_d02ecd52cd3e.slice/crio-63be27c43d48d47990866c59a6277eaaf72bce264a2033dd182dd98df5a0e35c WatchSource:0}: Error finding container 63be27c43d48d47990866c59a6277eaaf72bce264a2033dd182dd98df5a0e35c: Status 404 returned error can't find the container with id 63be27c43d48d47990866c59a6277eaaf72bce264a2033dd182dd98df5a0e35c Mar 16 00:23:28 crc kubenswrapper[4751]: I0316 00:23:28.705849 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"af16bec7-f32d-4931-b196-d02ecd52cd3e","Type":"ContainerStarted","Data":"63be27c43d48d47990866c59a6277eaaf72bce264a2033dd182dd98df5a0e35c"} Mar 16 00:23:30 crc kubenswrapper[4751]: I0316 00:23:30.454964 4751 scope.go:117] "RemoveContainer" containerID="fc969b536c76c025ccca9dfa9fbdec2a9b5478db1e634c4a3ce7f89d2e997033" Mar 16 00:23:30 crc kubenswrapper[4751]: I0316 00:23:30.455370 4751 scope.go:117] "RemoveContainer" containerID="73382b4d68e09937b679d0d640ec13f0225a86850a1abf61fce61f8528294b4e" Mar 16 00:23:31 crc kubenswrapper[4751]: I0316 00:23:31.454853 4751 scope.go:117] "RemoveContainer" containerID="bfd7b74e7f8cb0313d6cdbc197a083cc75f2c4a2f1eaa292c92dce0968d3205f" Mar 16 00:23:31 crc kubenswrapper[4751]: I0316 00:23:31.728361 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp" event={"ID":"24d89dab-af8f-4823-b6a4-37e9fa41a466","Type":"ContainerStarted","Data":"d350a4aa8be94f2fd97e4705e9c0469de007c825196403dc0dd678209d7b93fe"} Mar 16 00:23:31 crc kubenswrapper[4751]: I0316 00:23:31.732223 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq" event={"ID":"93427807-2b70-4cbe-9e6e-6fc6486a0020","Type":"ContainerStarted","Data":"e88bb6dc1c5b1d26613da73c2a31d07a92d3e8216778a86f0b911f988e03d4a2"} Mar 16 00:23:32 crc kubenswrapper[4751]: I0316 00:23:32.742134 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk" event={"ID":"ccf828f2-8f37-48da-8f95-b0e0ce744f34","Type":"ContainerStarted","Data":"7b04f77fbab9da6247a167d28413ad59e8de294acd42216c41490fd02a916ec0"} Mar 16 00:23:34 crc kubenswrapper[4751]: I0316 00:23:34.455255 4751 scope.go:117] "RemoveContainer" containerID="3b04832e7795cbb8ad926bac73ce87de233de3a8e525b1a119d18ee4c68ebde8" Mar 16 00:23:35 crc kubenswrapper[4751]: I0316 00:23:35.455006 4751 scope.go:117] "RemoveContainer" containerID="caf2db675fa55e2ad6e938a5a36435079ade14e6fe942460c7fdafdc8f6cb3c5" Mar 16 00:23:41 crc kubenswrapper[4751]: I0316 00:23:41.832187 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6" event={"ID":"3229d246-5e60-4ae7-8d1a-0fcf116d4038","Type":"ContainerStarted","Data":"b627ad39a0d4f8eb75fad3709e51e1f7ac21bc202d6a452cc9c278c0098e36d3"} Mar 16 00:23:41 crc kubenswrapper[4751]: I0316 00:23:41.835255 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4" event={"ID":"87bbd13c-ea3b-4664-bafe-92ac339d62ef","Type":"ContainerStarted","Data":"2527fd006e55da313ff9249e9bf0071f7ebed6ff851d836456407dc2433c8a7c"} Mar 16 00:23:41 crc kubenswrapper[4751]: I0316 00:23:41.838248 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"af16bec7-f32d-4931-b196-d02ecd52cd3e","Type":"ContainerStarted","Data":"de584900d7d7917c94fc0e079ce7a5ccd7bde18200ef38db9c1f1ac670f0e29b"} Mar 16 00:23:41 crc kubenswrapper[4751]: I0316 00:23:41.923137 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=2.069352872 podStartE2EDuration="14.923082871s" podCreationTimestamp="2026-03-16 00:23:27 +0000 UTC" firstStartedPulling="2026-03-16 00:23:28.228707046 +0000 UTC m=+1036.498039356" lastFinishedPulling="2026-03-16 00:23:41.082437025 +0000 UTC m=+1049.351769355" observedRunningTime="2026-03-16 00:23:41.910081966 +0000 UTC m=+1050.179414356" watchObservedRunningTime="2026-03-16 00:23:41.923082871 +0000 UTC m=+1050.192415201" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.289140 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-wglf5"] Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.290333 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.293987 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.294131 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.294226 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.294392 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.294493 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.294568 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.318503 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-wglf5"] Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.403856 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-config\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.403913 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-publisher\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.403944 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.403965 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lntrc\" (UniqueName: \"kubernetes.io/projected/14b84395-4c38-4c95-bd6a-a9033759e580-kube-api-access-lntrc\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.404002 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.404313 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-sensubility-config\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.404384 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-healthcheck-log\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.505735 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.505835 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-sensubility-config\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.505866 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-healthcheck-log\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.505898 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-config\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.505940 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-publisher\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.505981 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.506007 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lntrc\" (UniqueName: \"kubernetes.io/projected/14b84395-4c38-4c95-bd6a-a9033759e580-kube-api-access-lntrc\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.506796 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-healthcheck-log\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.507060 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.507200 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.507408 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-publisher\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.507544 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-config\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.508144 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-sensubility-config\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.540821 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lntrc\" (UniqueName: \"kubernetes.io/projected/14b84395-4c38-4c95-bd6a-a9033759e580-kube-api-access-lntrc\") pod \"stf-smoketest-smoke1-wglf5\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.616540 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.648745 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.649474 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.659522 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.810955 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qfb5\" (UniqueName: \"kubernetes.io/projected/2fa941aa-2828-4979-a0d3-22235c375ff3-kube-api-access-5qfb5\") pod \"curl\" (UID: \"2fa941aa-2828-4979-a0d3-22235c375ff3\") " pod="service-telemetry/curl" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.889780 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-wglf5"] Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.912707 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qfb5\" (UniqueName: \"kubernetes.io/projected/2fa941aa-2828-4979-a0d3-22235c375ff3-kube-api-access-5qfb5\") pod \"curl\" (UID: \"2fa941aa-2828-4979-a0d3-22235c375ff3\") " pod="service-telemetry/curl" Mar 16 00:23:42 crc kubenswrapper[4751]: I0316 00:23:42.931933 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qfb5\" (UniqueName: \"kubernetes.io/projected/2fa941aa-2828-4979-a0d3-22235c375ff3-kube-api-access-5qfb5\") pod \"curl\" (UID: \"2fa941aa-2828-4979-a0d3-22235c375ff3\") " pod="service-telemetry/curl" Mar 16 00:23:43 crc kubenswrapper[4751]: I0316 00:23:43.039926 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 16 00:23:43 crc kubenswrapper[4751]: I0316 00:23:43.283246 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 16 00:23:43 crc kubenswrapper[4751]: W0316 00:23:43.290885 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fa941aa_2828_4979_a0d3_22235c375ff3.slice/crio-2eb82f84f5638cc0830e2c112dbeb8dd8a5310c194604affc120d9837b60a1f2 WatchSource:0}: Error finding container 2eb82f84f5638cc0830e2c112dbeb8dd8a5310c194604affc120d9837b60a1f2: Status 404 returned error can't find the container with id 2eb82f84f5638cc0830e2c112dbeb8dd8a5310c194604affc120d9837b60a1f2 Mar 16 00:23:43 crc kubenswrapper[4751]: I0316 00:23:43.855526 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"2fa941aa-2828-4979-a0d3-22235c375ff3","Type":"ContainerStarted","Data":"2eb82f84f5638cc0830e2c112dbeb8dd8a5310c194604affc120d9837b60a1f2"} Mar 16 00:23:43 crc kubenswrapper[4751]: I0316 00:23:43.856928 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-wglf5" event={"ID":"14b84395-4c38-4c95-bd6a-a9033759e580","Type":"ContainerStarted","Data":"cc4d81e0c126b7f678bed88267daba13fb73ba24180244ff9e3ecefcad0dc0f9"} Mar 16 00:23:46 crc kubenswrapper[4751]: I0316 00:23:46.885263 4751 generic.go:334] "Generic (PLEG): container finished" podID="2fa941aa-2828-4979-a0d3-22235c375ff3" containerID="ac4b58bf62943dcd12c4be01d426dc6a5d390826fe20f84e4a77b9e257ec4f21" exitCode=0 Mar 16 00:23:46 crc kubenswrapper[4751]: I0316 00:23:46.885358 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"2fa941aa-2828-4979-a0d3-22235c375ff3","Type":"ContainerDied","Data":"ac4b58bf62943dcd12c4be01d426dc6a5d390826fe20f84e4a77b9e257ec4f21"} Mar 16 00:23:49 crc kubenswrapper[4751]: I0316 00:23:49.630744 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 16 00:23:49 crc kubenswrapper[4751]: I0316 00:23:49.768283 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_2fa941aa-2828-4979-a0d3-22235c375ff3/curl/0.log" Mar 16 00:23:49 crc kubenswrapper[4751]: I0316 00:23:49.818575 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qfb5\" (UniqueName: \"kubernetes.io/projected/2fa941aa-2828-4979-a0d3-22235c375ff3-kube-api-access-5qfb5\") pod \"2fa941aa-2828-4979-a0d3-22235c375ff3\" (UID: \"2fa941aa-2828-4979-a0d3-22235c375ff3\") " Mar 16 00:23:49 crc kubenswrapper[4751]: I0316 00:23:49.825309 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fa941aa-2828-4979-a0d3-22235c375ff3-kube-api-access-5qfb5" (OuterVolumeSpecName: "kube-api-access-5qfb5") pod "2fa941aa-2828-4979-a0d3-22235c375ff3" (UID: "2fa941aa-2828-4979-a0d3-22235c375ff3"). InnerVolumeSpecName "kube-api-access-5qfb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:23:49 crc kubenswrapper[4751]: I0316 00:23:49.920897 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qfb5\" (UniqueName: \"kubernetes.io/projected/2fa941aa-2828-4979-a0d3-22235c375ff3-kube-api-access-5qfb5\") on node \"crc\" DevicePath \"\"" Mar 16 00:23:49 crc kubenswrapper[4751]: I0316 00:23:49.932356 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"2fa941aa-2828-4979-a0d3-22235c375ff3","Type":"ContainerDied","Data":"2eb82f84f5638cc0830e2c112dbeb8dd8a5310c194604affc120d9837b60a1f2"} Mar 16 00:23:49 crc kubenswrapper[4751]: I0316 00:23:49.932434 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2eb82f84f5638cc0830e2c112dbeb8dd8a5310c194604affc120d9837b60a1f2" Mar 16 00:23:49 crc kubenswrapper[4751]: I0316 00:23:49.932395 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 16 00:23:50 crc kubenswrapper[4751]: I0316 00:23:50.043379 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-78bcbbdcff-pg68n_ec8ad70c-3602-46da-8cd2-b344f5365a71/prometheus-webhook-snmp/0.log" Mar 16 00:23:54 crc kubenswrapper[4751]: I0316 00:23:54.989176 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-wglf5" event={"ID":"14b84395-4c38-4c95-bd6a-a9033759e580","Type":"ContainerStarted","Data":"526b999c4402b241605f72a55a5a5b3d7a6d819bddb3575aad5542e396f23706"} Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.136099 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560344-d5j4b"] Mar 16 00:24:00 crc kubenswrapper[4751]: E0316 00:24:00.137447 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa941aa-2828-4979-a0d3-22235c375ff3" containerName="curl" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.137463 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa941aa-2828-4979-a0d3-22235c375ff3" containerName="curl" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.137615 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fa941aa-2828-4979-a0d3-22235c375ff3" containerName="curl" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.138306 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560344-d5j4b" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.142040 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.142556 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.142972 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.150090 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560344-d5j4b"] Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.187750 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtwrn\" (UniqueName: \"kubernetes.io/projected/5682a1ec-e3e2-47f2-b820-9b34062bba55-kube-api-access-dtwrn\") pod \"auto-csr-approver-29560344-d5j4b\" (UID: \"5682a1ec-e3e2-47f2-b820-9b34062bba55\") " pod="openshift-infra/auto-csr-approver-29560344-d5j4b" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.289206 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtwrn\" (UniqueName: \"kubernetes.io/projected/5682a1ec-e3e2-47f2-b820-9b34062bba55-kube-api-access-dtwrn\") pod \"auto-csr-approver-29560344-d5j4b\" (UID: \"5682a1ec-e3e2-47f2-b820-9b34062bba55\") " pod="openshift-infra/auto-csr-approver-29560344-d5j4b" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.310397 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtwrn\" (UniqueName: \"kubernetes.io/projected/5682a1ec-e3e2-47f2-b820-9b34062bba55-kube-api-access-dtwrn\") pod \"auto-csr-approver-29560344-d5j4b\" (UID: \"5682a1ec-e3e2-47f2-b820-9b34062bba55\") " pod="openshift-infra/auto-csr-approver-29560344-d5j4b" Mar 16 00:24:00 crc kubenswrapper[4751]: I0316 00:24:00.465860 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560344-d5j4b" Mar 16 00:24:02 crc kubenswrapper[4751]: I0316 00:24:02.274261 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560344-d5j4b"] Mar 16 00:24:02 crc kubenswrapper[4751]: W0316 00:24:02.280251 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5682a1ec_e3e2_47f2_b820_9b34062bba55.slice/crio-7bb85619679f74ac3eed402849a7793d85fe4cdc5b1c42e1272e6284e0c396b2 WatchSource:0}: Error finding container 7bb85619679f74ac3eed402849a7793d85fe4cdc5b1c42e1272e6284e0c396b2: Status 404 returned error can't find the container with id 7bb85619679f74ac3eed402849a7793d85fe4cdc5b1c42e1272e6284e0c396b2 Mar 16 00:24:03 crc kubenswrapper[4751]: I0316 00:24:03.082214 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-wglf5" event={"ID":"14b84395-4c38-4c95-bd6a-a9033759e580","Type":"ContainerStarted","Data":"1ec1d6e9ac9ab20bf4dba9f99b573004512d9cc459f28b59a58698a48b3b764a"} Mar 16 00:24:03 crc kubenswrapper[4751]: I0316 00:24:03.083628 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560344-d5j4b" event={"ID":"5682a1ec-e3e2-47f2-b820-9b34062bba55","Type":"ContainerStarted","Data":"7bb85619679f74ac3eed402849a7793d85fe4cdc5b1c42e1272e6284e0c396b2"} Mar 16 00:24:03 crc kubenswrapper[4751]: I0316 00:24:03.113624 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-wglf5" podStartSLOduration=1.960528573 podStartE2EDuration="21.113600898s" podCreationTimestamp="2026-03-16 00:23:42 +0000 UTC" firstStartedPulling="2026-03-16 00:23:42.914864395 +0000 UTC m=+1051.184196695" lastFinishedPulling="2026-03-16 00:24:02.06793672 +0000 UTC m=+1070.337269020" observedRunningTime="2026-03-16 00:24:03.103319419 +0000 UTC m=+1071.372651719" watchObservedRunningTime="2026-03-16 00:24:03.113600898 +0000 UTC m=+1071.382933198" Mar 16 00:24:04 crc kubenswrapper[4751]: I0316 00:24:04.096143 4751 generic.go:334] "Generic (PLEG): container finished" podID="5682a1ec-e3e2-47f2-b820-9b34062bba55" containerID="613123c92a5390f5761e5ba9d788f48b48a01cf99f429953be8f09dbbe984ebb" exitCode=0 Mar 16 00:24:04 crc kubenswrapper[4751]: I0316 00:24:04.098834 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560344-d5j4b" event={"ID":"5682a1ec-e3e2-47f2-b820-9b34062bba55","Type":"ContainerDied","Data":"613123c92a5390f5761e5ba9d788f48b48a01cf99f429953be8f09dbbe984ebb"} Mar 16 00:24:05 crc kubenswrapper[4751]: I0316 00:24:05.507999 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560344-d5j4b" Mar 16 00:24:05 crc kubenswrapper[4751]: I0316 00:24:05.690325 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtwrn\" (UniqueName: \"kubernetes.io/projected/5682a1ec-e3e2-47f2-b820-9b34062bba55-kube-api-access-dtwrn\") pod \"5682a1ec-e3e2-47f2-b820-9b34062bba55\" (UID: \"5682a1ec-e3e2-47f2-b820-9b34062bba55\") " Mar 16 00:24:05 crc kubenswrapper[4751]: I0316 00:24:05.695672 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5682a1ec-e3e2-47f2-b820-9b34062bba55-kube-api-access-dtwrn" (OuterVolumeSpecName: "kube-api-access-dtwrn") pod "5682a1ec-e3e2-47f2-b820-9b34062bba55" (UID: "5682a1ec-e3e2-47f2-b820-9b34062bba55"). InnerVolumeSpecName "kube-api-access-dtwrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:24:05 crc kubenswrapper[4751]: I0316 00:24:05.794059 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtwrn\" (UniqueName: \"kubernetes.io/projected/5682a1ec-e3e2-47f2-b820-9b34062bba55-kube-api-access-dtwrn\") on node \"crc\" DevicePath \"\"" Mar 16 00:24:06 crc kubenswrapper[4751]: I0316 00:24:06.113821 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560344-d5j4b" event={"ID":"5682a1ec-e3e2-47f2-b820-9b34062bba55","Type":"ContainerDied","Data":"7bb85619679f74ac3eed402849a7793d85fe4cdc5b1c42e1272e6284e0c396b2"} Mar 16 00:24:06 crc kubenswrapper[4751]: I0316 00:24:06.113856 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bb85619679f74ac3eed402849a7793d85fe4cdc5b1c42e1272e6284e0c396b2" Mar 16 00:24:06 crc kubenswrapper[4751]: I0316 00:24:06.113950 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560344-d5j4b" Mar 16 00:24:06 crc kubenswrapper[4751]: I0316 00:24:06.595546 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29560338-ghpdr"] Mar 16 00:24:06 crc kubenswrapper[4751]: I0316 00:24:06.605878 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29560338-ghpdr"] Mar 16 00:24:08 crc kubenswrapper[4751]: I0316 00:24:08.468676 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2b5d3b3-849a-4d38-b757-fe5f36befaa5" path="/var/lib/kubelet/pods/a2b5d3b3-849a-4d38-b757-fe5f36befaa5/volumes" Mar 16 00:24:17 crc kubenswrapper[4751]: I0316 00:24:17.778621 4751 scope.go:117] "RemoveContainer" containerID="89df761e05427df967335e82e301b66d2caf73c0911761cb993a5e767380b188" Mar 16 00:24:20 crc kubenswrapper[4751]: I0316 00:24:20.211743 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-78bcbbdcff-pg68n_ec8ad70c-3602-46da-8cd2-b344f5365a71/prometheus-webhook-snmp/0.log" Mar 16 00:24:27 crc kubenswrapper[4751]: I0316 00:24:27.854720 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:24:27 crc kubenswrapper[4751]: I0316 00:24:27.856855 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:24:28 crc kubenswrapper[4751]: I0316 00:24:28.306248 4751 generic.go:334] "Generic (PLEG): container finished" podID="14b84395-4c38-4c95-bd6a-a9033759e580" containerID="526b999c4402b241605f72a55a5a5b3d7a6d819bddb3575aad5542e396f23706" exitCode=0 Mar 16 00:24:28 crc kubenswrapper[4751]: I0316 00:24:28.306368 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-wglf5" event={"ID":"14b84395-4c38-4c95-bd6a-a9033759e580","Type":"ContainerDied","Data":"526b999c4402b241605f72a55a5a5b3d7a6d819bddb3575aad5542e396f23706"} Mar 16 00:24:28 crc kubenswrapper[4751]: I0316 00:24:28.307512 4751 scope.go:117] "RemoveContainer" containerID="526b999c4402b241605f72a55a5a5b3d7a6d819bddb3575aad5542e396f23706" Mar 16 00:24:34 crc kubenswrapper[4751]: I0316 00:24:34.364901 4751 generic.go:334] "Generic (PLEG): container finished" podID="14b84395-4c38-4c95-bd6a-a9033759e580" containerID="1ec1d6e9ac9ab20bf4dba9f99b573004512d9cc459f28b59a58698a48b3b764a" exitCode=0 Mar 16 00:24:34 crc kubenswrapper[4751]: I0316 00:24:34.364971 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-wglf5" event={"ID":"14b84395-4c38-4c95-bd6a-a9033759e580","Type":"ContainerDied","Data":"1ec1d6e9ac9ab20bf4dba9f99b573004512d9cc459f28b59a58698a48b3b764a"} Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.740078 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.907647 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-entrypoint-script\") pod \"14b84395-4c38-4c95-bd6a-a9033759e580\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.908046 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-publisher\") pod \"14b84395-4c38-4c95-bd6a-a9033759e580\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.908086 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-config\") pod \"14b84395-4c38-4c95-bd6a-a9033759e580\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.908155 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-entrypoint-script\") pod \"14b84395-4c38-4c95-bd6a-a9033759e580\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.908191 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-healthcheck-log\") pod \"14b84395-4c38-4c95-bd6a-a9033759e580\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.908261 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lntrc\" (UniqueName: \"kubernetes.io/projected/14b84395-4c38-4c95-bd6a-a9033759e580-kube-api-access-lntrc\") pod \"14b84395-4c38-4c95-bd6a-a9033759e580\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.908325 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-sensubility-config\") pod \"14b84395-4c38-4c95-bd6a-a9033759e580\" (UID: \"14b84395-4c38-4c95-bd6a-a9033759e580\") " Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.914325 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14b84395-4c38-4c95-bd6a-a9033759e580-kube-api-access-lntrc" (OuterVolumeSpecName: "kube-api-access-lntrc") pod "14b84395-4c38-4c95-bd6a-a9033759e580" (UID: "14b84395-4c38-4c95-bd6a-a9033759e580"). InnerVolumeSpecName "kube-api-access-lntrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.939871 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "14b84395-4c38-4c95-bd6a-a9033759e580" (UID: "14b84395-4c38-4c95-bd6a-a9033759e580"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.943788 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "14b84395-4c38-4c95-bd6a-a9033759e580" (UID: "14b84395-4c38-4c95-bd6a-a9033759e580"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.944507 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "14b84395-4c38-4c95-bd6a-a9033759e580" (UID: "14b84395-4c38-4c95-bd6a-a9033759e580"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.947003 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "14b84395-4c38-4c95-bd6a-a9033759e580" (UID: "14b84395-4c38-4c95-bd6a-a9033759e580"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.948932 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "14b84395-4c38-4c95-bd6a-a9033759e580" (UID: "14b84395-4c38-4c95-bd6a-a9033759e580"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:24:35 crc kubenswrapper[4751]: I0316 00:24:35.949056 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "14b84395-4c38-4c95-bd6a-a9033759e580" (UID: "14b84395-4c38-4c95-bd6a-a9033759e580"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.009794 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lntrc\" (UniqueName: \"kubernetes.io/projected/14b84395-4c38-4c95-bd6a-a9033759e580-kube-api-access-lntrc\") on node \"crc\" DevicePath \"\"" Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.009836 4751 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-sensubility-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.009854 4751 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.009872 4751 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.009887 4751 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-collectd-config\") on node \"crc\" DevicePath \"\"" Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.009937 4751 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.009960 4751 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/14b84395-4c38-4c95-bd6a-a9033759e580-healthcheck-log\") on node \"crc\" DevicePath \"\"" Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.389045 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-wglf5" event={"ID":"14b84395-4c38-4c95-bd6a-a9033759e580","Type":"ContainerDied","Data":"cc4d81e0c126b7f678bed88267daba13fb73ba24180244ff9e3ecefcad0dc0f9"} Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.389206 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc4d81e0c126b7f678bed88267daba13fb73ba24180244ff9e3ecefcad0dc0f9" Mar 16 00:24:36 crc kubenswrapper[4751]: I0316 00:24:36.389325 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-wglf5" Mar 16 00:24:37 crc kubenswrapper[4751]: I0316 00:24:37.790563 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-wglf5_14b84395-4c38-4c95-bd6a-a9033759e580/smoketest-collectd/0.log" Mar 16 00:24:38 crc kubenswrapper[4751]: I0316 00:24:38.018878 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-wglf5_14b84395-4c38-4c95-bd6a-a9033759e580/smoketest-ceilometer/0.log" Mar 16 00:24:38 crc kubenswrapper[4751]: I0316 00:24:38.296869 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-68864d46cb-2trgn_4054f75c-cc02-418d-a301-e3a7c92ef630/default-interconnect/0.log" Mar 16 00:24:38 crc kubenswrapper[4751]: I0316 00:24:38.515712 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq_93427807-2b70-4cbe-9e6e-6fc6486a0020/bridge/2.log" Mar 16 00:24:38 crc kubenswrapper[4751]: I0316 00:24:38.723993 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7996dc9458-4z2lq_93427807-2b70-4cbe-9e6e-6fc6486a0020/sg-core/0.log" Mar 16 00:24:38 crc kubenswrapper[4751]: I0316 00:24:38.964527 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4_87bbd13c-ea3b-4664-bafe-92ac339d62ef/bridge/2.log" Mar 16 00:24:39 crc kubenswrapper[4751]: I0316 00:24:39.218676 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-75d776b5b4-x6tq4_87bbd13c-ea3b-4664-bafe-92ac339d62ef/sg-core/0.log" Mar 16 00:24:39 crc kubenswrapper[4751]: I0316 00:24:39.462000 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk_ccf828f2-8f37-48da-8f95-b0e0ce744f34/bridge/2.log" Mar 16 00:24:39 crc kubenswrapper[4751]: I0316 00:24:39.686738 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-b57f974ff-5l6xk_ccf828f2-8f37-48da-8f95-b0e0ce744f34/sg-core/0.log" Mar 16 00:24:39 crc kubenswrapper[4751]: I0316 00:24:39.945610 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6_3229d246-5e60-4ae7-8d1a-0fcf116d4038/bridge/2.log" Mar 16 00:24:40 crc kubenswrapper[4751]: I0316 00:24:40.182994 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-8656d6cbcc-475c6_3229d246-5e60-4ae7-8d1a-0fcf116d4038/sg-core/0.log" Mar 16 00:24:40 crc kubenswrapper[4751]: I0316 00:24:40.464413 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp_24d89dab-af8f-4823-b6a4-37e9fa41a466/bridge/2.log" Mar 16 00:24:40 crc kubenswrapper[4751]: I0316 00:24:40.740014 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-6864f4fb65-df9gp_24d89dab-af8f-4823-b6a4-37e9fa41a466/sg-core/0.log" Mar 16 00:24:43 crc kubenswrapper[4751]: I0316 00:24:43.532393 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-697d54f584-rdntd_6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7/operator/0.log" Mar 16 00:24:43 crc kubenswrapper[4751]: I0316 00:24:43.773333 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_aaf54795-c77b-42bf-a088-5104d905ab2d/prometheus/0.log" Mar 16 00:24:44 crc kubenswrapper[4751]: I0316 00:24:44.026915 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_f222ef67-630e-4203-8857-b2d4fb4f1e89/elasticsearch/0.log" Mar 16 00:24:44 crc kubenswrapper[4751]: I0316 00:24:44.266942 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-78bcbbdcff-pg68n_ec8ad70c-3602-46da-8cd2-b344f5365a71/prometheus-webhook-snmp/0.log" Mar 16 00:24:44 crc kubenswrapper[4751]: I0316 00:24:44.509913 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_32d5f19b-e735-49bd-8212-3367fd97cc49/alertmanager/0.log" Mar 16 00:24:57 crc kubenswrapper[4751]: I0316 00:24:57.460144 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-fcbbbcfcf-ngbch_2d362a30-4cc0-43f3-bae1-482bcc3ad4e5/operator/0.log" Mar 16 00:24:57 crc kubenswrapper[4751]: I0316 00:24:57.854091 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:24:57 crc kubenswrapper[4751]: I0316 00:24:57.854274 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:25:00 crc kubenswrapper[4751]: I0316 00:25:00.118975 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-697d54f584-rdntd_6b438b99-00f6-497f-bfcc-6b2c1d6e0fa7/operator/0.log" Mar 16 00:25:00 crc kubenswrapper[4751]: I0316 00:25:00.354234 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_af16bec7-f32d-4931-b196-d02ecd52cd3e/qdr/0.log" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.746882 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6gvhq/must-gather-wzmck"] Mar 16 00:25:25 crc kubenswrapper[4751]: E0316 00:25:25.748890 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14b84395-4c38-4c95-bd6a-a9033759e580" containerName="smoketest-collectd" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.749017 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="14b84395-4c38-4c95-bd6a-a9033759e580" containerName="smoketest-collectd" Mar 16 00:25:25 crc kubenswrapper[4751]: E0316 00:25:25.749128 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14b84395-4c38-4c95-bd6a-a9033759e580" containerName="smoketest-ceilometer" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.749230 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="14b84395-4c38-4c95-bd6a-a9033759e580" containerName="smoketest-ceilometer" Mar 16 00:25:25 crc kubenswrapper[4751]: E0316 00:25:25.749323 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5682a1ec-e3e2-47f2-b820-9b34062bba55" containerName="oc" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.749410 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5682a1ec-e3e2-47f2-b820-9b34062bba55" containerName="oc" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.749679 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5682a1ec-e3e2-47f2-b820-9b34062bba55" containerName="oc" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.749777 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="14b84395-4c38-4c95-bd6a-a9033759e580" containerName="smoketest-collectd" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.749869 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="14b84395-4c38-4c95-bd6a-a9033759e580" containerName="smoketest-ceilometer" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.750806 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.753830 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6gvhq"/"kube-root-ca.crt" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.759642 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6gvhq"/"openshift-service-ca.crt" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.768850 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6gvhq/must-gather-wzmck"] Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.908587 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2f947f9a-4484-4500-b30c-73ac450df3c8-must-gather-output\") pod \"must-gather-wzmck\" (UID: \"2f947f9a-4484-4500-b30c-73ac450df3c8\") " pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:25:25 crc kubenswrapper[4751]: I0316 00:25:25.908650 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmlgl\" (UniqueName: \"kubernetes.io/projected/2f947f9a-4484-4500-b30c-73ac450df3c8-kube-api-access-lmlgl\") pod \"must-gather-wzmck\" (UID: \"2f947f9a-4484-4500-b30c-73ac450df3c8\") " pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:25:26 crc kubenswrapper[4751]: I0316 00:25:26.009582 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2f947f9a-4484-4500-b30c-73ac450df3c8-must-gather-output\") pod \"must-gather-wzmck\" (UID: \"2f947f9a-4484-4500-b30c-73ac450df3c8\") " pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:25:26 crc kubenswrapper[4751]: I0316 00:25:26.009647 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmlgl\" (UniqueName: \"kubernetes.io/projected/2f947f9a-4484-4500-b30c-73ac450df3c8-kube-api-access-lmlgl\") pod \"must-gather-wzmck\" (UID: \"2f947f9a-4484-4500-b30c-73ac450df3c8\") " pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:25:26 crc kubenswrapper[4751]: I0316 00:25:26.010304 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2f947f9a-4484-4500-b30c-73ac450df3c8-must-gather-output\") pod \"must-gather-wzmck\" (UID: \"2f947f9a-4484-4500-b30c-73ac450df3c8\") " pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:25:26 crc kubenswrapper[4751]: I0316 00:25:26.027416 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmlgl\" (UniqueName: \"kubernetes.io/projected/2f947f9a-4484-4500-b30c-73ac450df3c8-kube-api-access-lmlgl\") pod \"must-gather-wzmck\" (UID: \"2f947f9a-4484-4500-b30c-73ac450df3c8\") " pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:25:26 crc kubenswrapper[4751]: I0316 00:25:26.075372 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:25:26 crc kubenswrapper[4751]: I0316 00:25:26.285834 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6gvhq/must-gather-wzmck"] Mar 16 00:25:26 crc kubenswrapper[4751]: W0316 00:25:26.292652 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f947f9a_4484_4500_b30c_73ac450df3c8.slice/crio-2f72e8c8ccb3040e8b7f1fd0cbc0f6200debf494054211c288a3489c6dff0f5b WatchSource:0}: Error finding container 2f72e8c8ccb3040e8b7f1fd0cbc0f6200debf494054211c288a3489c6dff0f5b: Status 404 returned error can't find the container with id 2f72e8c8ccb3040e8b7f1fd0cbc0f6200debf494054211c288a3489c6dff0f5b Mar 16 00:25:26 crc kubenswrapper[4751]: I0316 00:25:26.295261 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 16 00:25:26 crc kubenswrapper[4751]: I0316 00:25:26.853128 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6gvhq/must-gather-wzmck" event={"ID":"2f947f9a-4484-4500-b30c-73ac450df3c8","Type":"ContainerStarted","Data":"2f72e8c8ccb3040e8b7f1fd0cbc0f6200debf494054211c288a3489c6dff0f5b"} Mar 16 00:25:27 crc kubenswrapper[4751]: I0316 00:25:27.854497 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:25:27 crc kubenswrapper[4751]: I0316 00:25:27.854583 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:25:27 crc kubenswrapper[4751]: I0316 00:25:27.854655 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:25:27 crc kubenswrapper[4751]: I0316 00:25:27.855731 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e0586c280086ba5e014ebfa12b8cdaadcbcd7999bb69b5913e321ce4b81f9a1"} pod="openshift-machine-config-operator/machine-config-daemon-6fccz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 16 00:25:27 crc kubenswrapper[4751]: I0316 00:25:27.855798 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" containerID="cri-o://8e0586c280086ba5e014ebfa12b8cdaadcbcd7999bb69b5913e321ce4b81f9a1" gracePeriod=600 Mar 16 00:25:28 crc kubenswrapper[4751]: I0316 00:25:28.873683 4751 generic.go:334] "Generic (PLEG): container finished" podID="25429719-d18d-4909-8455-701d09423c36" containerID="8e0586c280086ba5e014ebfa12b8cdaadcbcd7999bb69b5913e321ce4b81f9a1" exitCode=0 Mar 16 00:25:28 crc kubenswrapper[4751]: I0316 00:25:28.873759 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerDied","Data":"8e0586c280086ba5e014ebfa12b8cdaadcbcd7999bb69b5913e321ce4b81f9a1"} Mar 16 00:25:28 crc kubenswrapper[4751]: I0316 00:25:28.874497 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerStarted","Data":"b254c97eba1dcf7ddcb5c7c13381976881e3d0a12fdb91f3d8cf634a8127de36"} Mar 16 00:25:28 crc kubenswrapper[4751]: I0316 00:25:28.874537 4751 scope.go:117] "RemoveContainer" containerID="65efad03d473120c12153253332bfeb08a8e6a82c87fa93d5e003a10153418a4" Mar 16 00:25:33 crc kubenswrapper[4751]: I0316 00:25:33.916993 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6gvhq/must-gather-wzmck" event={"ID":"2f947f9a-4484-4500-b30c-73ac450df3c8","Type":"ContainerStarted","Data":"34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061"} Mar 16 00:25:33 crc kubenswrapper[4751]: I0316 00:25:33.917623 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6gvhq/must-gather-wzmck" event={"ID":"2f947f9a-4484-4500-b30c-73ac450df3c8","Type":"ContainerStarted","Data":"808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58"} Mar 16 00:25:33 crc kubenswrapper[4751]: I0316 00:25:33.948371 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6gvhq/must-gather-wzmck" podStartSLOduration=2.516912746 podStartE2EDuration="8.948350631s" podCreationTimestamp="2026-03-16 00:25:25 +0000 UTC" firstStartedPulling="2026-03-16 00:25:26.295202779 +0000 UTC m=+1154.564535069" lastFinishedPulling="2026-03-16 00:25:32.726640664 +0000 UTC m=+1160.995972954" observedRunningTime="2026-03-16 00:25:33.941987942 +0000 UTC m=+1162.211320242" watchObservedRunningTime="2026-03-16 00:25:33.948350631 +0000 UTC m=+1162.217682931" Mar 16 00:25:40 crc kubenswrapper[4751]: I0316 00:25:40.942337 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-g8q4w"] Mar 16 00:25:40 crc kubenswrapper[4751]: I0316 00:25:40.949279 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:40 crc kubenswrapper[4751]: I0316 00:25:40.964080 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-g8q4w"] Mar 16 00:25:41 crc kubenswrapper[4751]: I0316 00:25:41.034701 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcfc9\" (UniqueName: \"kubernetes.io/projected/2c5e5b41-7cd0-4fae-92da-b60923e195ed-kube-api-access-fcfc9\") pod \"infrawatch-operators-g8q4w\" (UID: \"2c5e5b41-7cd0-4fae-92da-b60923e195ed\") " pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:41 crc kubenswrapper[4751]: I0316 00:25:41.136213 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcfc9\" (UniqueName: \"kubernetes.io/projected/2c5e5b41-7cd0-4fae-92da-b60923e195ed-kube-api-access-fcfc9\") pod \"infrawatch-operators-g8q4w\" (UID: \"2c5e5b41-7cd0-4fae-92da-b60923e195ed\") " pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:41 crc kubenswrapper[4751]: I0316 00:25:41.163045 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcfc9\" (UniqueName: \"kubernetes.io/projected/2c5e5b41-7cd0-4fae-92da-b60923e195ed-kube-api-access-fcfc9\") pod \"infrawatch-operators-g8q4w\" (UID: \"2c5e5b41-7cd0-4fae-92da-b60923e195ed\") " pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:41 crc kubenswrapper[4751]: I0316 00:25:41.276514 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:41 crc kubenswrapper[4751]: I0316 00:25:41.556412 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-g8q4w"] Mar 16 00:25:41 crc kubenswrapper[4751]: I0316 00:25:41.974544 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g8q4w" event={"ID":"2c5e5b41-7cd0-4fae-92da-b60923e195ed","Type":"ContainerStarted","Data":"3b55a9b3f4624c193db1117b140ebcf95163258b4cc068b957290b9b4198d437"} Mar 16 00:25:41 crc kubenswrapper[4751]: I0316 00:25:41.975074 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g8q4w" event={"ID":"2c5e5b41-7cd0-4fae-92da-b60923e195ed","Type":"ContainerStarted","Data":"c6b68af5798494b2035553842c82eb31a02e5bbc9b461e332c4c893e114c7ba6"} Mar 16 00:25:41 crc kubenswrapper[4751]: I0316 00:25:41.998777 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-g8q4w" podStartSLOduration=1.868625969 podStartE2EDuration="1.998757027s" podCreationTimestamp="2026-03-16 00:25:40 +0000 UTC" firstStartedPulling="2026-03-16 00:25:41.564694287 +0000 UTC m=+1169.834026577" lastFinishedPulling="2026-03-16 00:25:41.694825335 +0000 UTC m=+1169.964157635" observedRunningTime="2026-03-16 00:25:41.992599184 +0000 UTC m=+1170.261931484" watchObservedRunningTime="2026-03-16 00:25:41.998757027 +0000 UTC m=+1170.268089327" Mar 16 00:25:51 crc kubenswrapper[4751]: I0316 00:25:51.277745 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:51 crc kubenswrapper[4751]: I0316 00:25:51.278521 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:51 crc kubenswrapper[4751]: I0316 00:25:51.326263 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:52 crc kubenswrapper[4751]: I0316 00:25:52.101221 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:54 crc kubenswrapper[4751]: I0316 00:25:54.914034 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-g8q4w"] Mar 16 00:25:54 crc kubenswrapper[4751]: I0316 00:25:54.914317 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-g8q4w" podUID="2c5e5b41-7cd0-4fae-92da-b60923e195ed" containerName="registry-server" containerID="cri-o://3b55a9b3f4624c193db1117b140ebcf95163258b4cc068b957290b9b4198d437" gracePeriod=2 Mar 16 00:25:55 crc kubenswrapper[4751]: I0316 00:25:55.102220 4751 generic.go:334] "Generic (PLEG): container finished" podID="2c5e5b41-7cd0-4fae-92da-b60923e195ed" containerID="3b55a9b3f4624c193db1117b140ebcf95163258b4cc068b957290b9b4198d437" exitCode=0 Mar 16 00:25:55 crc kubenswrapper[4751]: I0316 00:25:55.102301 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g8q4w" event={"ID":"2c5e5b41-7cd0-4fae-92da-b60923e195ed","Type":"ContainerDied","Data":"3b55a9b3f4624c193db1117b140ebcf95163258b4cc068b957290b9b4198d437"} Mar 16 00:25:55 crc kubenswrapper[4751]: I0316 00:25:55.306164 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:55 crc kubenswrapper[4751]: I0316 00:25:55.448639 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcfc9\" (UniqueName: \"kubernetes.io/projected/2c5e5b41-7cd0-4fae-92da-b60923e195ed-kube-api-access-fcfc9\") pod \"2c5e5b41-7cd0-4fae-92da-b60923e195ed\" (UID: \"2c5e5b41-7cd0-4fae-92da-b60923e195ed\") " Mar 16 00:25:55 crc kubenswrapper[4751]: I0316 00:25:55.468877 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c5e5b41-7cd0-4fae-92da-b60923e195ed-kube-api-access-fcfc9" (OuterVolumeSpecName: "kube-api-access-fcfc9") pod "2c5e5b41-7cd0-4fae-92da-b60923e195ed" (UID: "2c5e5b41-7cd0-4fae-92da-b60923e195ed"). InnerVolumeSpecName "kube-api-access-fcfc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:25:55 crc kubenswrapper[4751]: I0316 00:25:55.551183 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcfc9\" (UniqueName: \"kubernetes.io/projected/2c5e5b41-7cd0-4fae-92da-b60923e195ed-kube-api-access-fcfc9\") on node \"crc\" DevicePath \"\"" Mar 16 00:25:56 crc kubenswrapper[4751]: I0316 00:25:56.114704 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g8q4w" event={"ID":"2c5e5b41-7cd0-4fae-92da-b60923e195ed","Type":"ContainerDied","Data":"c6b68af5798494b2035553842c82eb31a02e5bbc9b461e332c4c893e114c7ba6"} Mar 16 00:25:56 crc kubenswrapper[4751]: I0316 00:25:56.114773 4751 scope.go:117] "RemoveContainer" containerID="3b55a9b3f4624c193db1117b140ebcf95163258b4cc068b957290b9b4198d437" Mar 16 00:25:56 crc kubenswrapper[4751]: I0316 00:25:56.114915 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g8q4w" Mar 16 00:25:56 crc kubenswrapper[4751]: I0316 00:25:56.161470 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-g8q4w"] Mar 16 00:25:56 crc kubenswrapper[4751]: I0316 00:25:56.167917 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-g8q4w"] Mar 16 00:25:56 crc kubenswrapper[4751]: I0316 00:25:56.462721 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c5e5b41-7cd0-4fae-92da-b60923e195ed" path="/var/lib/kubelet/pods/2c5e5b41-7cd0-4fae-92da-b60923e195ed/volumes" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.141567 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560346-6hrkd"] Mar 16 00:26:00 crc kubenswrapper[4751]: E0316 00:26:00.142493 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c5e5b41-7cd0-4fae-92da-b60923e195ed" containerName="registry-server" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.142509 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c5e5b41-7cd0-4fae-92da-b60923e195ed" containerName="registry-server" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.142676 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c5e5b41-7cd0-4fae-92da-b60923e195ed" containerName="registry-server" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.143311 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560346-6hrkd" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.145922 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.146594 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.153249 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.153457 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560346-6hrkd"] Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.220393 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5shfl\" (UniqueName: \"kubernetes.io/projected/3946b9f4-7f0a-4c56-9dd8-35270fa93e95-kube-api-access-5shfl\") pod \"auto-csr-approver-29560346-6hrkd\" (UID: \"3946b9f4-7f0a-4c56-9dd8-35270fa93e95\") " pod="openshift-infra/auto-csr-approver-29560346-6hrkd" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.322509 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5shfl\" (UniqueName: \"kubernetes.io/projected/3946b9f4-7f0a-4c56-9dd8-35270fa93e95-kube-api-access-5shfl\") pod \"auto-csr-approver-29560346-6hrkd\" (UID: \"3946b9f4-7f0a-4c56-9dd8-35270fa93e95\") " pod="openshift-infra/auto-csr-approver-29560346-6hrkd" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.344188 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5shfl\" (UniqueName: \"kubernetes.io/projected/3946b9f4-7f0a-4c56-9dd8-35270fa93e95-kube-api-access-5shfl\") pod \"auto-csr-approver-29560346-6hrkd\" (UID: \"3946b9f4-7f0a-4c56-9dd8-35270fa93e95\") " pod="openshift-infra/auto-csr-approver-29560346-6hrkd" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.480631 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560346-6hrkd" Mar 16 00:26:00 crc kubenswrapper[4751]: I0316 00:26:00.736820 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560346-6hrkd"] Mar 16 00:26:01 crc kubenswrapper[4751]: I0316 00:26:01.163884 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560346-6hrkd" event={"ID":"3946b9f4-7f0a-4c56-9dd8-35270fa93e95","Type":"ContainerStarted","Data":"2a58a5d83f827a7711deba82b9d4f10af24fcd82d81dfe1a10630d5f1c9facda"} Mar 16 00:26:02 crc kubenswrapper[4751]: I0316 00:26:02.177332 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560346-6hrkd" event={"ID":"3946b9f4-7f0a-4c56-9dd8-35270fa93e95","Type":"ContainerStarted","Data":"6a33c4961b7336c9be6a8cc8bd983eecc6ac58be15ff72d5a1b29a3cebcfa9e5"} Mar 16 00:26:02 crc kubenswrapper[4751]: I0316 00:26:02.192464 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29560346-6hrkd" podStartSLOduration=1.155724771 podStartE2EDuration="2.192440498s" podCreationTimestamp="2026-03-16 00:26:00 +0000 UTC" firstStartedPulling="2026-03-16 00:26:00.748353821 +0000 UTC m=+1189.017686111" lastFinishedPulling="2026-03-16 00:26:01.785069538 +0000 UTC m=+1190.054401838" observedRunningTime="2026-03-16 00:26:02.19146688 +0000 UTC m=+1190.460799190" watchObservedRunningTime="2026-03-16 00:26:02.192440498 +0000 UTC m=+1190.461772828" Mar 16 00:26:03 crc kubenswrapper[4751]: I0316 00:26:03.188472 4751 generic.go:334] "Generic (PLEG): container finished" podID="3946b9f4-7f0a-4c56-9dd8-35270fa93e95" containerID="6a33c4961b7336c9be6a8cc8bd983eecc6ac58be15ff72d5a1b29a3cebcfa9e5" exitCode=0 Mar 16 00:26:03 crc kubenswrapper[4751]: I0316 00:26:03.188546 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560346-6hrkd" event={"ID":"3946b9f4-7f0a-4c56-9dd8-35270fa93e95","Type":"ContainerDied","Data":"6a33c4961b7336c9be6a8cc8bd983eecc6ac58be15ff72d5a1b29a3cebcfa9e5"} Mar 16 00:26:04 crc kubenswrapper[4751]: I0316 00:26:04.491532 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560346-6hrkd" Mar 16 00:26:04 crc kubenswrapper[4751]: I0316 00:26:04.590178 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5shfl\" (UniqueName: \"kubernetes.io/projected/3946b9f4-7f0a-4c56-9dd8-35270fa93e95-kube-api-access-5shfl\") pod \"3946b9f4-7f0a-4c56-9dd8-35270fa93e95\" (UID: \"3946b9f4-7f0a-4c56-9dd8-35270fa93e95\") " Mar 16 00:26:04 crc kubenswrapper[4751]: I0316 00:26:04.597143 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3946b9f4-7f0a-4c56-9dd8-35270fa93e95-kube-api-access-5shfl" (OuterVolumeSpecName: "kube-api-access-5shfl") pod "3946b9f4-7f0a-4c56-9dd8-35270fa93e95" (UID: "3946b9f4-7f0a-4c56-9dd8-35270fa93e95"). InnerVolumeSpecName "kube-api-access-5shfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:26:04 crc kubenswrapper[4751]: I0316 00:26:04.692256 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5shfl\" (UniqueName: \"kubernetes.io/projected/3946b9f4-7f0a-4c56-9dd8-35270fa93e95-kube-api-access-5shfl\") on node \"crc\" DevicePath \"\"" Mar 16 00:26:05 crc kubenswrapper[4751]: I0316 00:26:05.214387 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560346-6hrkd" event={"ID":"3946b9f4-7f0a-4c56-9dd8-35270fa93e95","Type":"ContainerDied","Data":"2a58a5d83f827a7711deba82b9d4f10af24fcd82d81dfe1a10630d5f1c9facda"} Mar 16 00:26:05 crc kubenswrapper[4751]: I0316 00:26:05.214448 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a58a5d83f827a7711deba82b9d4f10af24fcd82d81dfe1a10630d5f1c9facda" Mar 16 00:26:05 crc kubenswrapper[4751]: I0316 00:26:05.214512 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560346-6hrkd" Mar 16 00:26:05 crc kubenswrapper[4751]: I0316 00:26:05.255201 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29560340-bbd76"] Mar 16 00:26:05 crc kubenswrapper[4751]: I0316 00:26:05.261589 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29560340-bbd76"] Mar 16 00:26:06 crc kubenswrapper[4751]: I0316 00:26:06.467075 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc" path="/var/lib/kubelet/pods/3d7f31dc-1e3a-456c-b00c-d1f8d87e99fc/volumes" Mar 16 00:26:17 crc kubenswrapper[4751]: I0316 00:26:17.864754 4751 scope.go:117] "RemoveContainer" containerID="ef49699cabe449908208ec67e1ac110ada3edeb36c1b231b41f28a4c7af80c62" Mar 16 00:26:19 crc kubenswrapper[4751]: I0316 00:26:19.402318 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-w4zmm_5620fac3-59f5-49e4-9125-501353fbf836/control-plane-machine-set-operator/0.log" Mar 16 00:26:19 crc kubenswrapper[4751]: I0316 00:26:19.517736 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zb5z5_a932e9e3-ca99-4af3-a9e6-8888e31be60d/kube-rbac-proxy/0.log" Mar 16 00:26:19 crc kubenswrapper[4751]: I0316 00:26:19.563995 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zb5z5_a932e9e3-ca99-4af3-a9e6-8888e31be60d/machine-api-operator/0.log" Mar 16 00:26:32 crc kubenswrapper[4751]: I0316 00:26:32.861375 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-rv7zp_5c3f99b3-1665-4852-b3ab-e7b44e5593a4/cert-manager-controller/0.log" Mar 16 00:26:32 crc kubenswrapper[4751]: I0316 00:26:32.976814 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-wc58h_c3bc081d-6e79-44f0-bc4e-e3c30474aa2c/cert-manager-cainjector/0.log" Mar 16 00:26:33 crc kubenswrapper[4751]: I0316 00:26:33.000991 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-q9rzm_f9b31035-18b9-4a25-b230-7f448975ea02/cert-manager-webhook/0.log" Mar 16 00:26:47 crc kubenswrapper[4751]: I0316 00:26:47.122634 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-5t56t_9a21e689-2895-407e-8888-b5e66e9f69d8/prometheus-operator/0.log" Mar 16 00:26:47 crc kubenswrapper[4751]: I0316 00:26:47.211918 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp_91271a6a-d14a-4681-a311-58ddc1253908/prometheus-operator-admission-webhook/0.log" Mar 16 00:26:47 crc kubenswrapper[4751]: I0316 00:26:47.296048 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv_8d263a57-8030-4502-8dab-5fcce4c4f85e/prometheus-operator-admission-webhook/0.log" Mar 16 00:26:47 crc kubenswrapper[4751]: I0316 00:26:47.373133 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-dkpf2_2469430b-1d52-419e-b6c7-74be1d58ba4b/operator/0.log" Mar 16 00:26:47 crc kubenswrapper[4751]: I0316 00:26:47.468305 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-8bc59_c9b395e8-a7f9-4fee-b851-c4d382678e5c/perses-operator/0.log" Mar 16 00:27:02 crc kubenswrapper[4751]: I0316 00:27:02.716475 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz_6c930011-0311-4778-a108-885ea6ecd7ec/util/0.log" Mar 16 00:27:02 crc kubenswrapper[4751]: I0316 00:27:02.867730 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz_6c930011-0311-4778-a108-885ea6ecd7ec/pull/0.log" Mar 16 00:27:02 crc kubenswrapper[4751]: I0316 00:27:02.880838 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz_6c930011-0311-4778-a108-885ea6ecd7ec/pull/0.log" Mar 16 00:27:02 crc kubenswrapper[4751]: I0316 00:27:02.916016 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz_6c930011-0311-4778-a108-885ea6ecd7ec/util/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.081464 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz_6c930011-0311-4778-a108-885ea6ecd7ec/pull/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.141733 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz_6c930011-0311-4778-a108-885ea6ecd7ec/util/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.162639 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fpx5pz_6c930011-0311-4778-a108-885ea6ecd7ec/extract/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.252208 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm_89cd9003-4d55-44e5-8f14-9caceb86e67a/util/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.553970 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm_89cd9003-4d55-44e5-8f14-9caceb86e67a/pull/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.557505 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm_89cd9003-4d55-44e5-8f14-9caceb86e67a/pull/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.591235 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm_89cd9003-4d55-44e5-8f14-9caceb86e67a/util/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.790287 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm_89cd9003-4d55-44e5-8f14-9caceb86e67a/util/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.791338 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm_89cd9003-4d55-44e5-8f14-9caceb86e67a/extract/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.794275 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39evrhpm_89cd9003-4d55-44e5-8f14-9caceb86e67a/pull/0.log" Mar 16 00:27:03 crc kubenswrapper[4751]: I0316 00:27:03.990250 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts_d54c0437-6747-476e-a69c-b7b59ee0bd7d/util/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.127875 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts_d54c0437-6747-476e-a69c-b7b59ee0bd7d/util/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.205838 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts_d54c0437-6747-476e-a69c-b7b59ee0bd7d/pull/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.215779 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts_d54c0437-6747-476e-a69c-b7b59ee0bd7d/pull/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.401519 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts_d54c0437-6747-476e-a69c-b7b59ee0bd7d/extract/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.402763 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts_d54c0437-6747-476e-a69c-b7b59ee0bd7d/util/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.440518 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e586fts_d54c0437-6747-476e-a69c-b7b59ee0bd7d/pull/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.581976 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4_bd8ea216-286a-4575-8b3e-fd8c3f2afcda/util/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.764456 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4_bd8ea216-286a-4575-8b3e-fd8c3f2afcda/util/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.782695 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4_bd8ea216-286a-4575-8b3e-fd8c3f2afcda/pull/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.818656 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4_bd8ea216-286a-4575-8b3e-fd8c3f2afcda/pull/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.981870 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4_bd8ea216-286a-4575-8b3e-fd8c3f2afcda/pull/0.log" Mar 16 00:27:04 crc kubenswrapper[4751]: I0316 00:27:04.992290 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4_bd8ea216-286a-4575-8b3e-fd8c3f2afcda/extract/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.037735 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08t78k4_bd8ea216-286a-4575-8b3e-fd8c3f2afcda/util/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.170451 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-57pn4_76c15cfb-0b0f-4f9f-94eb-22b0d736f175/extract-utilities/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.346744 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-57pn4_76c15cfb-0b0f-4f9f-94eb-22b0d736f175/extract-content/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.377993 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-57pn4_76c15cfb-0b0f-4f9f-94eb-22b0d736f175/extract-utilities/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.401212 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-57pn4_76c15cfb-0b0f-4f9f-94eb-22b0d736f175/extract-content/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.573316 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-57pn4_76c15cfb-0b0f-4f9f-94eb-22b0d736f175/extract-utilities/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.576046 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-57pn4_76c15cfb-0b0f-4f9f-94eb-22b0d736f175/extract-content/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.788086 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nnkxp_647a9f72-750d-4c57-8684-4c841604a40e/extract-utilities/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.881063 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-57pn4_76c15cfb-0b0f-4f9f-94eb-22b0d736f175/registry-server/0.log" Mar 16 00:27:05 crc kubenswrapper[4751]: I0316 00:27:05.941742 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nnkxp_647a9f72-750d-4c57-8684-4c841604a40e/extract-content/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.144642 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nnkxp_647a9f72-750d-4c57-8684-4c841604a40e/extract-content/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.149321 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nnkxp_647a9f72-750d-4c57-8684-4c841604a40e/extract-utilities/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.312950 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nnkxp_647a9f72-750d-4c57-8684-4c841604a40e/extract-content/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.329778 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nnkxp_647a9f72-750d-4c57-8684-4c841604a40e/extract-utilities/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.358234 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ksmlj_9674dab9-cf73-46a0-8077-1ec770c4f73c/marketplace-operator/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.490677 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nnkxp_647a9f72-750d-4c57-8684-4c841604a40e/registry-server/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.491784 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j8n4q_1568ce0c-e5c5-4fa1-a95b-470216db436f/extract-utilities/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.649855 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j8n4q_1568ce0c-e5c5-4fa1-a95b-470216db436f/extract-utilities/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.676164 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j8n4q_1568ce0c-e5c5-4fa1-a95b-470216db436f/extract-content/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.681239 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j8n4q_1568ce0c-e5c5-4fa1-a95b-470216db436f/extract-content/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.811575 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j8n4q_1568ce0c-e5c5-4fa1-a95b-470216db436f/extract-content/0.log" Mar 16 00:27:06 crc kubenswrapper[4751]: I0316 00:27:06.822687 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j8n4q_1568ce0c-e5c5-4fa1-a95b-470216db436f/extract-utilities/0.log" Mar 16 00:27:07 crc kubenswrapper[4751]: I0316 00:27:07.089497 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j8n4q_1568ce0c-e5c5-4fa1-a95b-470216db436f/registry-server/0.log" Mar 16 00:27:20 crc kubenswrapper[4751]: I0316 00:27:20.987098 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6c78c89f86-kvqbp_91271a6a-d14a-4681-a311-58ddc1253908/prometheus-operator-admission-webhook/0.log" Mar 16 00:27:21 crc kubenswrapper[4751]: I0316 00:27:21.025534 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-5t56t_9a21e689-2895-407e-8888-b5e66e9f69d8/prometheus-operator/0.log" Mar 16 00:27:21 crc kubenswrapper[4751]: I0316 00:27:21.067561 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6c78c89f86-tlqsv_8d263a57-8030-4502-8dab-5fcce4c4f85e/prometheus-operator-admission-webhook/0.log" Mar 16 00:27:21 crc kubenswrapper[4751]: I0316 00:27:21.128347 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-dkpf2_2469430b-1d52-419e-b6c7-74be1d58ba4b/operator/0.log" Mar 16 00:27:21 crc kubenswrapper[4751]: I0316 00:27:21.177261 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-8bc59_c9b395e8-a7f9-4fee-b851-c4d382678e5c/perses-operator/0.log" Mar 16 00:27:57 crc kubenswrapper[4751]: I0316 00:27:57.855013 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:27:57 crc kubenswrapper[4751]: I0316 00:27:57.855902 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.171055 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560348-nvn4b"] Mar 16 00:28:00 crc kubenswrapper[4751]: E0316 00:28:00.171617 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3946b9f4-7f0a-4c56-9dd8-35270fa93e95" containerName="oc" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.171630 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3946b9f4-7f0a-4c56-9dd8-35270fa93e95" containerName="oc" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.171776 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3946b9f4-7f0a-4c56-9dd8-35270fa93e95" containerName="oc" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.172419 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560348-nvn4b" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.174360 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.175629 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.177759 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.189869 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560348-nvn4b"] Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.361355 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57vqg\" (UniqueName: \"kubernetes.io/projected/08436e07-c954-46a9-a2cc-2ec516473b5c-kube-api-access-57vqg\") pod \"auto-csr-approver-29560348-nvn4b\" (UID: \"08436e07-c954-46a9-a2cc-2ec516473b5c\") " pod="openshift-infra/auto-csr-approver-29560348-nvn4b" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.462896 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57vqg\" (UniqueName: \"kubernetes.io/projected/08436e07-c954-46a9-a2cc-2ec516473b5c-kube-api-access-57vqg\") pod \"auto-csr-approver-29560348-nvn4b\" (UID: \"08436e07-c954-46a9-a2cc-2ec516473b5c\") " pod="openshift-infra/auto-csr-approver-29560348-nvn4b" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.487587 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57vqg\" (UniqueName: \"kubernetes.io/projected/08436e07-c954-46a9-a2cc-2ec516473b5c-kube-api-access-57vqg\") pod \"auto-csr-approver-29560348-nvn4b\" (UID: \"08436e07-c954-46a9-a2cc-2ec516473b5c\") " pod="openshift-infra/auto-csr-approver-29560348-nvn4b" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.497564 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560348-nvn4b" Mar 16 00:28:00 crc kubenswrapper[4751]: I0316 00:28:00.752308 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560348-nvn4b"] Mar 16 00:28:01 crc kubenswrapper[4751]: I0316 00:28:01.184907 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560348-nvn4b" event={"ID":"08436e07-c954-46a9-a2cc-2ec516473b5c","Type":"ContainerStarted","Data":"4a67e93500c4de861359512bb7f55a0c28e639e9f788047543c27ead548da945"} Mar 16 00:28:03 crc kubenswrapper[4751]: I0316 00:28:03.202457 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560348-nvn4b" event={"ID":"08436e07-c954-46a9-a2cc-2ec516473b5c","Type":"ContainerStarted","Data":"d066b6fec75c9d213208e05e3d550b2146e562f3e1974af6302f1d296f56bcba"} Mar 16 00:28:03 crc kubenswrapper[4751]: I0316 00:28:03.216873 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29560348-nvn4b" podStartSLOduration=1.477053422 podStartE2EDuration="3.216850734s" podCreationTimestamp="2026-03-16 00:28:00 +0000 UTC" firstStartedPulling="2026-03-16 00:28:00.769886622 +0000 UTC m=+1309.039218952" lastFinishedPulling="2026-03-16 00:28:02.509683934 +0000 UTC m=+1310.779016264" observedRunningTime="2026-03-16 00:28:03.215785444 +0000 UTC m=+1311.485117744" watchObservedRunningTime="2026-03-16 00:28:03.216850734 +0000 UTC m=+1311.486183034" Mar 16 00:28:04 crc kubenswrapper[4751]: I0316 00:28:04.215862 4751 generic.go:334] "Generic (PLEG): container finished" podID="08436e07-c954-46a9-a2cc-2ec516473b5c" containerID="d066b6fec75c9d213208e05e3d550b2146e562f3e1974af6302f1d296f56bcba" exitCode=0 Mar 16 00:28:04 crc kubenswrapper[4751]: I0316 00:28:04.215919 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560348-nvn4b" event={"ID":"08436e07-c954-46a9-a2cc-2ec516473b5c","Type":"ContainerDied","Data":"d066b6fec75c9d213208e05e3d550b2146e562f3e1974af6302f1d296f56bcba"} Mar 16 00:28:05 crc kubenswrapper[4751]: I0316 00:28:05.559279 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560348-nvn4b" Mar 16 00:28:05 crc kubenswrapper[4751]: I0316 00:28:05.568815 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57vqg\" (UniqueName: \"kubernetes.io/projected/08436e07-c954-46a9-a2cc-2ec516473b5c-kube-api-access-57vqg\") pod \"08436e07-c954-46a9-a2cc-2ec516473b5c\" (UID: \"08436e07-c954-46a9-a2cc-2ec516473b5c\") " Mar 16 00:28:05 crc kubenswrapper[4751]: I0316 00:28:05.579195 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08436e07-c954-46a9-a2cc-2ec516473b5c-kube-api-access-57vqg" (OuterVolumeSpecName: "kube-api-access-57vqg") pod "08436e07-c954-46a9-a2cc-2ec516473b5c" (UID: "08436e07-c954-46a9-a2cc-2ec516473b5c"). InnerVolumeSpecName "kube-api-access-57vqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:28:05 crc kubenswrapper[4751]: I0316 00:28:05.670408 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57vqg\" (UniqueName: \"kubernetes.io/projected/08436e07-c954-46a9-a2cc-2ec516473b5c-kube-api-access-57vqg\") on node \"crc\" DevicePath \"\"" Mar 16 00:28:06 crc kubenswrapper[4751]: I0316 00:28:06.239058 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560348-nvn4b" event={"ID":"08436e07-c954-46a9-a2cc-2ec516473b5c","Type":"ContainerDied","Data":"4a67e93500c4de861359512bb7f55a0c28e639e9f788047543c27ead548da945"} Mar 16 00:28:06 crc kubenswrapper[4751]: I0316 00:28:06.239175 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a67e93500c4de861359512bb7f55a0c28e639e9f788047543c27ead548da945" Mar 16 00:28:06 crc kubenswrapper[4751]: I0316 00:28:06.239181 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560348-nvn4b" Mar 16 00:28:06 crc kubenswrapper[4751]: I0316 00:28:06.639673 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29560342-qpmsc"] Mar 16 00:28:06 crc kubenswrapper[4751]: I0316 00:28:06.646427 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29560342-qpmsc"] Mar 16 00:28:08 crc kubenswrapper[4751]: I0316 00:28:08.474345 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dcc0fca-049b-4dbe-9470-8e4b89bccfb6" path="/var/lib/kubelet/pods/6dcc0fca-049b-4dbe-9470-8e4b89bccfb6/volumes" Mar 16 00:28:10 crc kubenswrapper[4751]: I0316 00:28:10.295098 4751 generic.go:334] "Generic (PLEG): container finished" podID="2f947f9a-4484-4500-b30c-73ac450df3c8" containerID="808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58" exitCode=0 Mar 16 00:28:10 crc kubenswrapper[4751]: I0316 00:28:10.295199 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6gvhq/must-gather-wzmck" event={"ID":"2f947f9a-4484-4500-b30c-73ac450df3c8","Type":"ContainerDied","Data":"808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58"} Mar 16 00:28:10 crc kubenswrapper[4751]: I0316 00:28:10.295842 4751 scope.go:117] "RemoveContainer" containerID="808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58" Mar 16 00:28:11 crc kubenswrapper[4751]: I0316 00:28:11.192026 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6gvhq_must-gather-wzmck_2f947f9a-4484-4500-b30c-73ac450df3c8/gather/0.log" Mar 16 00:28:17 crc kubenswrapper[4751]: I0316 00:28:17.981122 4751 scope.go:117] "RemoveContainer" containerID="e8a67af31e0d598a8cc3f6536d0d51d0a1e24bddb2370c1a6cc411cb39bc04e8" Mar 16 00:28:18 crc kubenswrapper[4751]: I0316 00:28:18.523213 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6gvhq/must-gather-wzmck"] Mar 16 00:28:18 crc kubenswrapper[4751]: I0316 00:28:18.524196 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-6gvhq/must-gather-wzmck" podUID="2f947f9a-4484-4500-b30c-73ac450df3c8" containerName="copy" containerID="cri-o://34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061" gracePeriod=2 Mar 16 00:28:18 crc kubenswrapper[4751]: I0316 00:28:18.534722 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6gvhq/must-gather-wzmck"] Mar 16 00:28:18 crc kubenswrapper[4751]: I0316 00:28:18.936235 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6gvhq_must-gather-wzmck_2f947f9a-4484-4500-b30c-73ac450df3c8/copy/0.log" Mar 16 00:28:18 crc kubenswrapper[4751]: I0316 00:28:18.936766 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:28:18 crc kubenswrapper[4751]: I0316 00:28:18.990171 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmlgl\" (UniqueName: \"kubernetes.io/projected/2f947f9a-4484-4500-b30c-73ac450df3c8-kube-api-access-lmlgl\") pod \"2f947f9a-4484-4500-b30c-73ac450df3c8\" (UID: \"2f947f9a-4484-4500-b30c-73ac450df3c8\") " Mar 16 00:28:18 crc kubenswrapper[4751]: I0316 00:28:18.991061 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2f947f9a-4484-4500-b30c-73ac450df3c8-must-gather-output\") pod \"2f947f9a-4484-4500-b30c-73ac450df3c8\" (UID: \"2f947f9a-4484-4500-b30c-73ac450df3c8\") " Mar 16 00:28:18 crc kubenswrapper[4751]: I0316 00:28:18.997889 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f947f9a-4484-4500-b30c-73ac450df3c8-kube-api-access-lmlgl" (OuterVolumeSpecName: "kube-api-access-lmlgl") pod "2f947f9a-4484-4500-b30c-73ac450df3c8" (UID: "2f947f9a-4484-4500-b30c-73ac450df3c8"). InnerVolumeSpecName "kube-api-access-lmlgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.058149 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f947f9a-4484-4500-b30c-73ac450df3c8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2f947f9a-4484-4500-b30c-73ac450df3c8" (UID: "2f947f9a-4484-4500-b30c-73ac450df3c8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.092245 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmlgl\" (UniqueName: \"kubernetes.io/projected/2f947f9a-4484-4500-b30c-73ac450df3c8-kube-api-access-lmlgl\") on node \"crc\" DevicePath \"\"" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.092274 4751 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2f947f9a-4484-4500-b30c-73ac450df3c8-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.379376 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6gvhq_must-gather-wzmck_2f947f9a-4484-4500-b30c-73ac450df3c8/copy/0.log" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.379868 4751 generic.go:334] "Generic (PLEG): container finished" podID="2f947f9a-4484-4500-b30c-73ac450df3c8" containerID="34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061" exitCode=143 Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.379974 4751 scope.go:117] "RemoveContainer" containerID="34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.380190 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6gvhq/must-gather-wzmck" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.406268 4751 scope.go:117] "RemoveContainer" containerID="808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.465498 4751 scope.go:117] "RemoveContainer" containerID="34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061" Mar 16 00:28:19 crc kubenswrapper[4751]: E0316 00:28:19.466006 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061\": container with ID starting with 34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061 not found: ID does not exist" containerID="34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.466033 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061"} err="failed to get container status \"34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061\": rpc error: code = NotFound desc = could not find container \"34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061\": container with ID starting with 34c15e88f29e1d0bb30793fc9da0b6753bd8bce878033b4fbcfd489d9d885061 not found: ID does not exist" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.466052 4751 scope.go:117] "RemoveContainer" containerID="808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58" Mar 16 00:28:19 crc kubenswrapper[4751]: E0316 00:28:19.466366 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58\": container with ID starting with 808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58 not found: ID does not exist" containerID="808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58" Mar 16 00:28:19 crc kubenswrapper[4751]: I0316 00:28:19.466384 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58"} err="failed to get container status \"808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58\": rpc error: code = NotFound desc = could not find container \"808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58\": container with ID starting with 808824ca8f29918e0f2266b74c97d3daa6d8db79dfff7085593bc3c45c694c58 not found: ID does not exist" Mar 16 00:28:20 crc kubenswrapper[4751]: I0316 00:28:20.464211 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f947f9a-4484-4500-b30c-73ac450df3c8" path="/var/lib/kubelet/pods/2f947f9a-4484-4500-b30c-73ac450df3c8/volumes" Mar 16 00:28:27 crc kubenswrapper[4751]: I0316 00:28:27.854237 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:28:27 crc kubenswrapper[4751]: I0316 00:28:27.854948 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:28:57 crc kubenswrapper[4751]: I0316 00:28:57.854810 4751 patch_prober.go:28] interesting pod/machine-config-daemon-6fccz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 16 00:28:57 crc kubenswrapper[4751]: I0316 00:28:57.855696 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 16 00:28:57 crc kubenswrapper[4751]: I0316 00:28:57.855818 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" Mar 16 00:28:57 crc kubenswrapper[4751]: I0316 00:28:57.857023 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b254c97eba1dcf7ddcb5c7c13381976881e3d0a12fdb91f3d8cf634a8127de36"} pod="openshift-machine-config-operator/machine-config-daemon-6fccz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 16 00:28:57 crc kubenswrapper[4751]: I0316 00:28:57.857189 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" podUID="25429719-d18d-4909-8455-701d09423c36" containerName="machine-config-daemon" containerID="cri-o://b254c97eba1dcf7ddcb5c7c13381976881e3d0a12fdb91f3d8cf634a8127de36" gracePeriod=600 Mar 16 00:28:58 crc kubenswrapper[4751]: I0316 00:28:58.737070 4751 generic.go:334] "Generic (PLEG): container finished" podID="25429719-d18d-4909-8455-701d09423c36" containerID="b254c97eba1dcf7ddcb5c7c13381976881e3d0a12fdb91f3d8cf634a8127de36" exitCode=0 Mar 16 00:28:58 crc kubenswrapper[4751]: I0316 00:28:58.738006 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerDied","Data":"b254c97eba1dcf7ddcb5c7c13381976881e3d0a12fdb91f3d8cf634a8127de36"} Mar 16 00:28:58 crc kubenswrapper[4751]: I0316 00:28:58.738285 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6fccz" event={"ID":"25429719-d18d-4909-8455-701d09423c36","Type":"ContainerStarted","Data":"03931a71df766325f0e5d026bbd9963e40a49224b9bdee5d38e3bac1ae89b697"} Mar 16 00:28:58 crc kubenswrapper[4751]: I0316 00:28:58.738504 4751 scope.go:117] "RemoveContainer" containerID="8e0586c280086ba5e014ebfa12b8cdaadcbcd7999bb69b5913e321ce4b81f9a1" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.283607 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lccjx"] Mar 16 00:29:04 crc kubenswrapper[4751]: E0316 00:29:04.284518 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f947f9a-4484-4500-b30c-73ac450df3c8" containerName="copy" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.284533 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f947f9a-4484-4500-b30c-73ac450df3c8" containerName="copy" Mar 16 00:29:04 crc kubenswrapper[4751]: E0316 00:29:04.284550 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08436e07-c954-46a9-a2cc-2ec516473b5c" containerName="oc" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.284560 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="08436e07-c954-46a9-a2cc-2ec516473b5c" containerName="oc" Mar 16 00:29:04 crc kubenswrapper[4751]: E0316 00:29:04.284585 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f947f9a-4484-4500-b30c-73ac450df3c8" containerName="gather" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.284592 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f947f9a-4484-4500-b30c-73ac450df3c8" containerName="gather" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.284742 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="08436e07-c954-46a9-a2cc-2ec516473b5c" containerName="oc" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.284756 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f947f9a-4484-4500-b30c-73ac450df3c8" containerName="gather" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.284779 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f947f9a-4484-4500-b30c-73ac450df3c8" containerName="copy" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.286066 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.309833 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lccjx"] Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.442145 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2fcs\" (UniqueName: \"kubernetes.io/projected/cf54ac77-97c7-4b79-aa61-d784a585f122-kube-api-access-c2fcs\") pod \"certified-operators-lccjx\" (UID: \"cf54ac77-97c7-4b79-aa61-d784a585f122\") " pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.442204 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf54ac77-97c7-4b79-aa61-d784a585f122-catalog-content\") pod \"certified-operators-lccjx\" (UID: \"cf54ac77-97c7-4b79-aa61-d784a585f122\") " pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.442231 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf54ac77-97c7-4b79-aa61-d784a585f122-utilities\") pod \"certified-operators-lccjx\" (UID: \"cf54ac77-97c7-4b79-aa61-d784a585f122\") " pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.543680 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2fcs\" (UniqueName: \"kubernetes.io/projected/cf54ac77-97c7-4b79-aa61-d784a585f122-kube-api-access-c2fcs\") pod \"certified-operators-lccjx\" (UID: \"cf54ac77-97c7-4b79-aa61-d784a585f122\") " pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.543802 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf54ac77-97c7-4b79-aa61-d784a585f122-catalog-content\") pod \"certified-operators-lccjx\" (UID: \"cf54ac77-97c7-4b79-aa61-d784a585f122\") " pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.543854 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf54ac77-97c7-4b79-aa61-d784a585f122-utilities\") pod \"certified-operators-lccjx\" (UID: \"cf54ac77-97c7-4b79-aa61-d784a585f122\") " pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.544361 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf54ac77-97c7-4b79-aa61-d784a585f122-utilities\") pod \"certified-operators-lccjx\" (UID: \"cf54ac77-97c7-4b79-aa61-d784a585f122\") " pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.544521 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf54ac77-97c7-4b79-aa61-d784a585f122-catalog-content\") pod \"certified-operators-lccjx\" (UID: \"cf54ac77-97c7-4b79-aa61-d784a585f122\") " pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.580424 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2fcs\" (UniqueName: \"kubernetes.io/projected/cf54ac77-97c7-4b79-aa61-d784a585f122-kube-api-access-c2fcs\") pod \"certified-operators-lccjx\" (UID: \"cf54ac77-97c7-4b79-aa61-d784a585f122\") " pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:04 crc kubenswrapper[4751]: I0316 00:29:04.609797 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:05 crc kubenswrapper[4751]: I0316 00:29:05.114383 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lccjx"] Mar 16 00:29:05 crc kubenswrapper[4751]: W0316 00:29:05.118839 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf54ac77_97c7_4b79_aa61_d784a585f122.slice/crio-7044e6385969225ad7e11e111b07c5ad8bb9b8f5674904aa45899ee01ddca1ee WatchSource:0}: Error finding container 7044e6385969225ad7e11e111b07c5ad8bb9b8f5674904aa45899ee01ddca1ee: Status 404 returned error can't find the container with id 7044e6385969225ad7e11e111b07c5ad8bb9b8f5674904aa45899ee01ddca1ee Mar 16 00:29:05 crc kubenswrapper[4751]: I0316 00:29:05.811650 4751 generic.go:334] "Generic (PLEG): container finished" podID="cf54ac77-97c7-4b79-aa61-d784a585f122" containerID="2e492582ccbc9a907be17bb982644301cb64ea77486c55b8e390f18f6428ad77" exitCode=0 Mar 16 00:29:05 crc kubenswrapper[4751]: I0316 00:29:05.811751 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lccjx" event={"ID":"cf54ac77-97c7-4b79-aa61-d784a585f122","Type":"ContainerDied","Data":"2e492582ccbc9a907be17bb982644301cb64ea77486c55b8e390f18f6428ad77"} Mar 16 00:29:05 crc kubenswrapper[4751]: I0316 00:29:05.812036 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lccjx" event={"ID":"cf54ac77-97c7-4b79-aa61-d784a585f122","Type":"ContainerStarted","Data":"7044e6385969225ad7e11e111b07c5ad8bb9b8f5674904aa45899ee01ddca1ee"} Mar 16 00:29:10 crc kubenswrapper[4751]: I0316 00:29:10.853557 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lccjx" event={"ID":"cf54ac77-97c7-4b79-aa61-d784a585f122","Type":"ContainerStarted","Data":"f8d659ce2bf6d20047af2220b9403a31d7542fab4bd885c46ff90939ec43eaa8"} Mar 16 00:29:11 crc kubenswrapper[4751]: I0316 00:29:11.867239 4751 generic.go:334] "Generic (PLEG): container finished" podID="cf54ac77-97c7-4b79-aa61-d784a585f122" containerID="f8d659ce2bf6d20047af2220b9403a31d7542fab4bd885c46ff90939ec43eaa8" exitCode=0 Mar 16 00:29:11 crc kubenswrapper[4751]: I0316 00:29:11.867309 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lccjx" event={"ID":"cf54ac77-97c7-4b79-aa61-d784a585f122","Type":"ContainerDied","Data":"f8d659ce2bf6d20047af2220b9403a31d7542fab4bd885c46ff90939ec43eaa8"} Mar 16 00:29:12 crc kubenswrapper[4751]: I0316 00:29:12.878036 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lccjx" event={"ID":"cf54ac77-97c7-4b79-aa61-d784a585f122","Type":"ContainerStarted","Data":"e30d617ca77b4a0ba716200fe343a822ec6046efc30d662c78f322693409dcc4"} Mar 16 00:29:12 crc kubenswrapper[4751]: I0316 00:29:12.903514 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lccjx" podStartSLOduration=2.389943252 podStartE2EDuration="8.903488493s" podCreationTimestamp="2026-03-16 00:29:04 +0000 UTC" firstStartedPulling="2026-03-16 00:29:05.814440983 +0000 UTC m=+1374.083773313" lastFinishedPulling="2026-03-16 00:29:12.327986254 +0000 UTC m=+1380.597318554" observedRunningTime="2026-03-16 00:29:12.897313369 +0000 UTC m=+1381.166645679" watchObservedRunningTime="2026-03-16 00:29:12.903488493 +0000 UTC m=+1381.172820793" Mar 16 00:29:14 crc kubenswrapper[4751]: I0316 00:29:14.610547 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:14 crc kubenswrapper[4751]: I0316 00:29:14.611791 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:15 crc kubenswrapper[4751]: I0316 00:29:15.658439 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lccjx" podUID="cf54ac77-97c7-4b79-aa61-d784a585f122" containerName="registry-server" probeResult="failure" output=< Mar 16 00:29:15 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Mar 16 00:29:15 crc kubenswrapper[4751]: > Mar 16 00:29:24 crc kubenswrapper[4751]: I0316 00:29:24.680839 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:24 crc kubenswrapper[4751]: I0316 00:29:24.747401 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lccjx" Mar 16 00:29:24 crc kubenswrapper[4751]: I0316 00:29:24.861696 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lccjx"] Mar 16 00:29:24 crc kubenswrapper[4751]: I0316 00:29:24.926882 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-57pn4"] Mar 16 00:29:24 crc kubenswrapper[4751]: I0316 00:29:24.927226 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-57pn4" podUID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerName="registry-server" containerID="cri-o://87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1" gracePeriod=2 Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.340829 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.516807 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-catalog-content\") pod \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.516864 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-utilities\") pod \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.516898 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27k94\" (UniqueName: \"kubernetes.io/projected/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-kube-api-access-27k94\") pod \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\" (UID: \"76c15cfb-0b0f-4f9f-94eb-22b0d736f175\") " Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.517604 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-utilities" (OuterVolumeSpecName: "utilities") pod "76c15cfb-0b0f-4f9f-94eb-22b0d736f175" (UID: "76c15cfb-0b0f-4f9f-94eb-22b0d736f175"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.522649 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-kube-api-access-27k94" (OuterVolumeSpecName: "kube-api-access-27k94") pod "76c15cfb-0b0f-4f9f-94eb-22b0d736f175" (UID: "76c15cfb-0b0f-4f9f-94eb-22b0d736f175"). InnerVolumeSpecName "kube-api-access-27k94". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.575764 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76c15cfb-0b0f-4f9f-94eb-22b0d736f175" (UID: "76c15cfb-0b0f-4f9f-94eb-22b0d736f175"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.617868 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.617900 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:29:25 crc kubenswrapper[4751]: I0316 00:29:25.617910 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27k94\" (UniqueName: \"kubernetes.io/projected/76c15cfb-0b0f-4f9f-94eb-22b0d736f175-kube-api-access-27k94\") on node \"crc\" DevicePath \"\"" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.001052 4751 generic.go:334] "Generic (PLEG): container finished" podID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerID="87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1" exitCode=0 Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.001096 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57pn4" event={"ID":"76c15cfb-0b0f-4f9f-94eb-22b0d736f175","Type":"ContainerDied","Data":"87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1"} Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.001159 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57pn4" event={"ID":"76c15cfb-0b0f-4f9f-94eb-22b0d736f175","Type":"ContainerDied","Data":"dad2568a386f3923d70b0f03f2775b7409dcf80746d7e5e9bc0d874aded1a2a9"} Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.001177 4751 scope.go:117] "RemoveContainer" containerID="87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.001132 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57pn4" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.022184 4751 scope.go:117] "RemoveContainer" containerID="11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.026273 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-57pn4"] Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.031543 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-57pn4"] Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.039370 4751 scope.go:117] "RemoveContainer" containerID="e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.060795 4751 scope.go:117] "RemoveContainer" containerID="87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1" Mar 16 00:29:26 crc kubenswrapper[4751]: E0316 00:29:26.061323 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1\": container with ID starting with 87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1 not found: ID does not exist" containerID="87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.061372 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1"} err="failed to get container status \"87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1\": rpc error: code = NotFound desc = could not find container \"87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1\": container with ID starting with 87df035b4d3dc354632453c3b2cbf7f373b03cebff19b071435ff744846dc3a1 not found: ID does not exist" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.061400 4751 scope.go:117] "RemoveContainer" containerID="11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c" Mar 16 00:29:26 crc kubenswrapper[4751]: E0316 00:29:26.063710 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c\": container with ID starting with 11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c not found: ID does not exist" containerID="11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.063741 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c"} err="failed to get container status \"11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c\": rpc error: code = NotFound desc = could not find container \"11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c\": container with ID starting with 11a8242c764cf0e8a9ee76fa8420dc087bb6f26007984bcbd61c20874bd32b0c not found: ID does not exist" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.063763 4751 scope.go:117] "RemoveContainer" containerID="e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b" Mar 16 00:29:26 crc kubenswrapper[4751]: E0316 00:29:26.064068 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b\": container with ID starting with e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b not found: ID does not exist" containerID="e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.064139 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b"} err="failed to get container status \"e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b\": rpc error: code = NotFound desc = could not find container \"e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b\": container with ID starting with e520fc902cfa47cc618516d2ef1001143f811fc1f87fe8eb6dd92932813c5d0b not found: ID does not exist" Mar 16 00:29:26 crc kubenswrapper[4751]: I0316 00:29:26.462362 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" path="/var/lib/kubelet/pods/76c15cfb-0b0f-4f9f-94eb-22b0d736f175/volumes" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.068544 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d4rpw"] Mar 16 00:29:32 crc kubenswrapper[4751]: E0316 00:29:32.069439 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerName="extract-content" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.069462 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerName="extract-content" Mar 16 00:29:32 crc kubenswrapper[4751]: E0316 00:29:32.069483 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerName="extract-utilities" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.069494 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerName="extract-utilities" Mar 16 00:29:32 crc kubenswrapper[4751]: E0316 00:29:32.069516 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerName="registry-server" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.069527 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerName="registry-server" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.069720 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="76c15cfb-0b0f-4f9f-94eb-22b0d736f175" containerName="registry-server" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.071070 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.100953 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4rpw"] Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.217465 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-utilities\") pod \"redhat-operators-d4rpw\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.217507 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwb5r\" (UniqueName: \"kubernetes.io/projected/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-kube-api-access-rwb5r\") pod \"redhat-operators-d4rpw\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.217531 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-catalog-content\") pod \"redhat-operators-d4rpw\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.318967 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-catalog-content\") pod \"redhat-operators-d4rpw\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.319154 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-utilities\") pod \"redhat-operators-d4rpw\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.319182 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwb5r\" (UniqueName: \"kubernetes.io/projected/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-kube-api-access-rwb5r\") pod \"redhat-operators-d4rpw\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.319889 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-catalog-content\") pod \"redhat-operators-d4rpw\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.319899 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-utilities\") pod \"redhat-operators-d4rpw\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.341590 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwb5r\" (UniqueName: \"kubernetes.io/projected/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-kube-api-access-rwb5r\") pod \"redhat-operators-d4rpw\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.396050 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:32 crc kubenswrapper[4751]: I0316 00:29:32.670134 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4rpw"] Mar 16 00:29:32 crc kubenswrapper[4751]: W0316 00:29:32.680242 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc46a75f3_03ed_45c3_80d6_b0ceb34a0514.slice/crio-46e4b32fc56decdd989d7cecf0128397973c560d17595444037644e3132f709f WatchSource:0}: Error finding container 46e4b32fc56decdd989d7cecf0128397973c560d17595444037644e3132f709f: Status 404 returned error can't find the container with id 46e4b32fc56decdd989d7cecf0128397973c560d17595444037644e3132f709f Mar 16 00:29:33 crc kubenswrapper[4751]: I0316 00:29:33.060167 4751 generic.go:334] "Generic (PLEG): container finished" podID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerID="236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd" exitCode=0 Mar 16 00:29:33 crc kubenswrapper[4751]: I0316 00:29:33.060218 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4rpw" event={"ID":"c46a75f3-03ed-45c3-80d6-b0ceb34a0514","Type":"ContainerDied","Data":"236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd"} Mar 16 00:29:33 crc kubenswrapper[4751]: I0316 00:29:33.060246 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4rpw" event={"ID":"c46a75f3-03ed-45c3-80d6-b0ceb34a0514","Type":"ContainerStarted","Data":"46e4b32fc56decdd989d7cecf0128397973c560d17595444037644e3132f709f"} Mar 16 00:29:34 crc kubenswrapper[4751]: I0316 00:29:34.071538 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4rpw" event={"ID":"c46a75f3-03ed-45c3-80d6-b0ceb34a0514","Type":"ContainerStarted","Data":"7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce"} Mar 16 00:29:35 crc kubenswrapper[4751]: I0316 00:29:35.083652 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4rpw" event={"ID":"c46a75f3-03ed-45c3-80d6-b0ceb34a0514","Type":"ContainerDied","Data":"7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce"} Mar 16 00:29:35 crc kubenswrapper[4751]: I0316 00:29:35.083844 4751 generic.go:334] "Generic (PLEG): container finished" podID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerID="7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce" exitCode=0 Mar 16 00:29:36 crc kubenswrapper[4751]: I0316 00:29:36.094604 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4rpw" event={"ID":"c46a75f3-03ed-45c3-80d6-b0ceb34a0514","Type":"ContainerStarted","Data":"b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f"} Mar 16 00:29:36 crc kubenswrapper[4751]: I0316 00:29:36.129642 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d4rpw" podStartSLOduration=1.659889861 podStartE2EDuration="4.129616065s" podCreationTimestamp="2026-03-16 00:29:32 +0000 UTC" firstStartedPulling="2026-03-16 00:29:33.0616351 +0000 UTC m=+1401.330967420" lastFinishedPulling="2026-03-16 00:29:35.531361294 +0000 UTC m=+1403.800693624" observedRunningTime="2026-03-16 00:29:36.120161618 +0000 UTC m=+1404.389493958" watchObservedRunningTime="2026-03-16 00:29:36.129616065 +0000 UTC m=+1404.398948395" Mar 16 00:29:42 crc kubenswrapper[4751]: I0316 00:29:42.397012 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:42 crc kubenswrapper[4751]: I0316 00:29:42.397726 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:43 crc kubenswrapper[4751]: I0316 00:29:43.476870 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d4rpw" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerName="registry-server" probeResult="failure" output=< Mar 16 00:29:43 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Mar 16 00:29:43 crc kubenswrapper[4751]: > Mar 16 00:29:52 crc kubenswrapper[4751]: I0316 00:29:52.479254 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:52 crc kubenswrapper[4751]: I0316 00:29:52.564944 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:52 crc kubenswrapper[4751]: I0316 00:29:52.730233 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4rpw"] Mar 16 00:29:54 crc kubenswrapper[4751]: I0316 00:29:54.271426 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d4rpw" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerName="registry-server" containerID="cri-o://b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f" gracePeriod=2 Mar 16 00:29:54 crc kubenswrapper[4751]: I0316 00:29:54.757387 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:54 crc kubenswrapper[4751]: I0316 00:29:54.912920 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-catalog-content\") pod \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " Mar 16 00:29:54 crc kubenswrapper[4751]: I0316 00:29:54.913013 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwb5r\" (UniqueName: \"kubernetes.io/projected/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-kube-api-access-rwb5r\") pod \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " Mar 16 00:29:54 crc kubenswrapper[4751]: I0316 00:29:54.913189 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-utilities\") pod \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\" (UID: \"c46a75f3-03ed-45c3-80d6-b0ceb34a0514\") " Mar 16 00:29:54 crc kubenswrapper[4751]: I0316 00:29:54.914553 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-utilities" (OuterVolumeSpecName: "utilities") pod "c46a75f3-03ed-45c3-80d6-b0ceb34a0514" (UID: "c46a75f3-03ed-45c3-80d6-b0ceb34a0514"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:29:54 crc kubenswrapper[4751]: I0316 00:29:54.922546 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-kube-api-access-rwb5r" (OuterVolumeSpecName: "kube-api-access-rwb5r") pod "c46a75f3-03ed-45c3-80d6-b0ceb34a0514" (UID: "c46a75f3-03ed-45c3-80d6-b0ceb34a0514"). InnerVolumeSpecName "kube-api-access-rwb5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.015516 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-utilities\") on node \"crc\" DevicePath \"\"" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.015552 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwb5r\" (UniqueName: \"kubernetes.io/projected/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-kube-api-access-rwb5r\") on node \"crc\" DevicePath \"\"" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.124749 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c46a75f3-03ed-45c3-80d6-b0ceb34a0514" (UID: "c46a75f3-03ed-45c3-80d6-b0ceb34a0514"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.219133 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46a75f3-03ed-45c3-80d6-b0ceb34a0514-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.299743 4751 generic.go:334] "Generic (PLEG): container finished" podID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerID="b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f" exitCode=0 Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.299912 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4rpw" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.304366 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4rpw" event={"ID":"c46a75f3-03ed-45c3-80d6-b0ceb34a0514","Type":"ContainerDied","Data":"b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f"} Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.304461 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4rpw" event={"ID":"c46a75f3-03ed-45c3-80d6-b0ceb34a0514","Type":"ContainerDied","Data":"46e4b32fc56decdd989d7cecf0128397973c560d17595444037644e3132f709f"} Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.304497 4751 scope.go:117] "RemoveContainer" containerID="b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.362249 4751 scope.go:117] "RemoveContainer" containerID="7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.362943 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4rpw"] Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.374633 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d4rpw"] Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.393686 4751 scope.go:117] "RemoveContainer" containerID="236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.424793 4751 scope.go:117] "RemoveContainer" containerID="b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f" Mar 16 00:29:55 crc kubenswrapper[4751]: E0316 00:29:55.425422 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f\": container with ID starting with b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f not found: ID does not exist" containerID="b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.425468 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f"} err="failed to get container status \"b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f\": rpc error: code = NotFound desc = could not find container \"b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f\": container with ID starting with b1d4906fb05bbf6051c1e6d76a575d161068b36fc867dbb2ddb4fbef6f63382f not found: ID does not exist" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.425495 4751 scope.go:117] "RemoveContainer" containerID="7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce" Mar 16 00:29:55 crc kubenswrapper[4751]: E0316 00:29:55.426030 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce\": container with ID starting with 7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce not found: ID does not exist" containerID="7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.426142 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce"} err="failed to get container status \"7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce\": rpc error: code = NotFound desc = could not find container \"7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce\": container with ID starting with 7a4a54e9c04c14face077cbb6cb3d483b381b7a3b9151df583500f619604e9ce not found: ID does not exist" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.426196 4751 scope.go:117] "RemoveContainer" containerID="236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd" Mar 16 00:29:55 crc kubenswrapper[4751]: E0316 00:29:55.426619 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd\": container with ID starting with 236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd not found: ID does not exist" containerID="236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd" Mar 16 00:29:55 crc kubenswrapper[4751]: I0316 00:29:55.426665 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd"} err="failed to get container status \"236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd\": rpc error: code = NotFound desc = could not find container \"236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd\": container with ID starting with 236b38570077baeea771992f293a155570ebed7a9a2466f53fd0e20537c616dd not found: ID does not exist" Mar 16 00:29:56 crc kubenswrapper[4751]: I0316 00:29:56.472030 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" path="/var/lib/kubelet/pods/c46a75f3-03ed-45c3-80d6-b0ceb34a0514/volumes" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.160307 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29560350-dmhzw"] Mar 16 00:30:00 crc kubenswrapper[4751]: E0316 00:30:00.160903 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerName="registry-server" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.160919 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerName="registry-server" Mar 16 00:30:00 crc kubenswrapper[4751]: E0316 00:30:00.160944 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerName="extract-utilities" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.160953 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerName="extract-utilities" Mar 16 00:30:00 crc kubenswrapper[4751]: E0316 00:30:00.160962 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerName="extract-content" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.160973 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerName="extract-content" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.161162 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="c46a75f3-03ed-45c3-80d6-b0ceb34a0514" containerName="registry-server" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.161668 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560350-dmhzw" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.166664 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.166902 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d4jlj" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.167024 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.178478 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr"] Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.179428 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.183286 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.184017 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.197235 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560350-dmhzw"] Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.230435 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr"] Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.311843 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b37c51b-4003-4901-9e1d-046230e92b0a-config-volume\") pod \"collect-profiles-29560350-vdnqr\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.311907 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b37c51b-4003-4901-9e1d-046230e92b0a-secret-volume\") pod \"collect-profiles-29560350-vdnqr\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.311965 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p9nn\" (UniqueName: \"kubernetes.io/projected/6b37c51b-4003-4901-9e1d-046230e92b0a-kube-api-access-6p9nn\") pod \"collect-profiles-29560350-vdnqr\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.311983 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz7zc\" (UniqueName: \"kubernetes.io/projected/55c7867c-32f1-4ee1-b7cb-9214ba1fda77-kube-api-access-rz7zc\") pod \"auto-csr-approver-29560350-dmhzw\" (UID: \"55c7867c-32f1-4ee1-b7cb-9214ba1fda77\") " pod="openshift-infra/auto-csr-approver-29560350-dmhzw" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.414504 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p9nn\" (UniqueName: \"kubernetes.io/projected/6b37c51b-4003-4901-9e1d-046230e92b0a-kube-api-access-6p9nn\") pod \"collect-profiles-29560350-vdnqr\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.414584 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz7zc\" (UniqueName: \"kubernetes.io/projected/55c7867c-32f1-4ee1-b7cb-9214ba1fda77-kube-api-access-rz7zc\") pod \"auto-csr-approver-29560350-dmhzw\" (UID: \"55c7867c-32f1-4ee1-b7cb-9214ba1fda77\") " pod="openshift-infra/auto-csr-approver-29560350-dmhzw" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.414672 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b37c51b-4003-4901-9e1d-046230e92b0a-config-volume\") pod \"collect-profiles-29560350-vdnqr\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.414767 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b37c51b-4003-4901-9e1d-046230e92b0a-secret-volume\") pod \"collect-profiles-29560350-vdnqr\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.418696 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b37c51b-4003-4901-9e1d-046230e92b0a-config-volume\") pod \"collect-profiles-29560350-vdnqr\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.435199 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b37c51b-4003-4901-9e1d-046230e92b0a-secret-volume\") pod \"collect-profiles-29560350-vdnqr\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.436523 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p9nn\" (UniqueName: \"kubernetes.io/projected/6b37c51b-4003-4901-9e1d-046230e92b0a-kube-api-access-6p9nn\") pod \"collect-profiles-29560350-vdnqr\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.438799 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz7zc\" (UniqueName: \"kubernetes.io/projected/55c7867c-32f1-4ee1-b7cb-9214ba1fda77-kube-api-access-rz7zc\") pod \"auto-csr-approver-29560350-dmhzw\" (UID: \"55c7867c-32f1-4ee1-b7cb-9214ba1fda77\") " pod="openshift-infra/auto-csr-approver-29560350-dmhzw" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.493114 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560350-dmhzw" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.499722 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.780342 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29560350-dmhzw"] Mar 16 00:30:00 crc kubenswrapper[4751]: I0316 00:30:00.964855 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr"] Mar 16 00:30:00 crc kubenswrapper[4751]: W0316 00:30:00.966909 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b37c51b_4003_4901_9e1d_046230e92b0a.slice/crio-fc678af32f423b62d0bd19a35df1ac2ccb60e19d8308898270db5560e70a5abd WatchSource:0}: Error finding container fc678af32f423b62d0bd19a35df1ac2ccb60e19d8308898270db5560e70a5abd: Status 404 returned error can't find the container with id fc678af32f423b62d0bd19a35df1ac2ccb60e19d8308898270db5560e70a5abd Mar 16 00:30:01 crc kubenswrapper[4751]: I0316 00:30:01.392063 4751 generic.go:334] "Generic (PLEG): container finished" podID="6b37c51b-4003-4901-9e1d-046230e92b0a" containerID="3681f3261948c44732002b688def2a18546a5a078adc2bfeb82084017f0c20a1" exitCode=0 Mar 16 00:30:01 crc kubenswrapper[4751]: I0316 00:30:01.392133 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" event={"ID":"6b37c51b-4003-4901-9e1d-046230e92b0a","Type":"ContainerDied","Data":"3681f3261948c44732002b688def2a18546a5a078adc2bfeb82084017f0c20a1"} Mar 16 00:30:01 crc kubenswrapper[4751]: I0316 00:30:01.392196 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" event={"ID":"6b37c51b-4003-4901-9e1d-046230e92b0a","Type":"ContainerStarted","Data":"fc678af32f423b62d0bd19a35df1ac2ccb60e19d8308898270db5560e70a5abd"} Mar 16 00:30:01 crc kubenswrapper[4751]: I0316 00:30:01.394789 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560350-dmhzw" event={"ID":"55c7867c-32f1-4ee1-b7cb-9214ba1fda77","Type":"ContainerStarted","Data":"a731686455e6c2894d25e2250888e7f028cddef885b53e3d6e210ad424144884"} Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.764992 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.849793 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b37c51b-4003-4901-9e1d-046230e92b0a-secret-volume\") pod \"6b37c51b-4003-4901-9e1d-046230e92b0a\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.849886 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p9nn\" (UniqueName: \"kubernetes.io/projected/6b37c51b-4003-4901-9e1d-046230e92b0a-kube-api-access-6p9nn\") pod \"6b37c51b-4003-4901-9e1d-046230e92b0a\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.849951 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b37c51b-4003-4901-9e1d-046230e92b0a-config-volume\") pod \"6b37c51b-4003-4901-9e1d-046230e92b0a\" (UID: \"6b37c51b-4003-4901-9e1d-046230e92b0a\") " Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.851042 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b37c51b-4003-4901-9e1d-046230e92b0a-config-volume" (OuterVolumeSpecName: "config-volume") pod "6b37c51b-4003-4901-9e1d-046230e92b0a" (UID: "6b37c51b-4003-4901-9e1d-046230e92b0a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.857782 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b37c51b-4003-4901-9e1d-046230e92b0a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6b37c51b-4003-4901-9e1d-046230e92b0a" (UID: "6b37c51b-4003-4901-9e1d-046230e92b0a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.857931 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b37c51b-4003-4901-9e1d-046230e92b0a-kube-api-access-6p9nn" (OuterVolumeSpecName: "kube-api-access-6p9nn") pod "6b37c51b-4003-4901-9e1d-046230e92b0a" (UID: "6b37c51b-4003-4901-9e1d-046230e92b0a"). InnerVolumeSpecName "kube-api-access-6p9nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.951623 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b37c51b-4003-4901-9e1d-046230e92b0a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.951673 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b37c51b-4003-4901-9e1d-046230e92b0a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 16 00:30:02 crc kubenswrapper[4751]: I0316 00:30:02.951693 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p9nn\" (UniqueName: \"kubernetes.io/projected/6b37c51b-4003-4901-9e1d-046230e92b0a-kube-api-access-6p9nn\") on node \"crc\" DevicePath \"\"" Mar 16 00:30:03 crc kubenswrapper[4751]: I0316 00:30:03.419808 4751 generic.go:334] "Generic (PLEG): container finished" podID="55c7867c-32f1-4ee1-b7cb-9214ba1fda77" containerID="59c17517495021266af09c24bad2af4f6d3a0b7ee74b1b870cc673261b954a1f" exitCode=0 Mar 16 00:30:03 crc kubenswrapper[4751]: I0316 00:30:03.420308 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560350-dmhzw" event={"ID":"55c7867c-32f1-4ee1-b7cb-9214ba1fda77","Type":"ContainerDied","Data":"59c17517495021266af09c24bad2af4f6d3a0b7ee74b1b870cc673261b954a1f"} Mar 16 00:30:03 crc kubenswrapper[4751]: I0316 00:30:03.424878 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" event={"ID":"6b37c51b-4003-4901-9e1d-046230e92b0a","Type":"ContainerDied","Data":"fc678af32f423b62d0bd19a35df1ac2ccb60e19d8308898270db5560e70a5abd"} Mar 16 00:30:03 crc kubenswrapper[4751]: I0316 00:30:03.424945 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc678af32f423b62d0bd19a35df1ac2ccb60e19d8308898270db5560e70a5abd" Mar 16 00:30:03 crc kubenswrapper[4751]: I0316 00:30:03.424964 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29560350-vdnqr" Mar 16 00:30:04 crc kubenswrapper[4751]: I0316 00:30:04.780216 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560350-dmhzw" Mar 16 00:30:04 crc kubenswrapper[4751]: I0316 00:30:04.884012 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz7zc\" (UniqueName: \"kubernetes.io/projected/55c7867c-32f1-4ee1-b7cb-9214ba1fda77-kube-api-access-rz7zc\") pod \"55c7867c-32f1-4ee1-b7cb-9214ba1fda77\" (UID: \"55c7867c-32f1-4ee1-b7cb-9214ba1fda77\") " Mar 16 00:30:04 crc kubenswrapper[4751]: I0316 00:30:04.892309 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55c7867c-32f1-4ee1-b7cb-9214ba1fda77-kube-api-access-rz7zc" (OuterVolumeSpecName: "kube-api-access-rz7zc") pod "55c7867c-32f1-4ee1-b7cb-9214ba1fda77" (UID: "55c7867c-32f1-4ee1-b7cb-9214ba1fda77"). InnerVolumeSpecName "kube-api-access-rz7zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 16 00:30:04 crc kubenswrapper[4751]: I0316 00:30:04.986755 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz7zc\" (UniqueName: \"kubernetes.io/projected/55c7867c-32f1-4ee1-b7cb-9214ba1fda77-kube-api-access-rz7zc\") on node \"crc\" DevicePath \"\"" Mar 16 00:30:05 crc kubenswrapper[4751]: I0316 00:30:05.454800 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29560350-dmhzw" event={"ID":"55c7867c-32f1-4ee1-b7cb-9214ba1fda77","Type":"ContainerDied","Data":"a731686455e6c2894d25e2250888e7f028cddef885b53e3d6e210ad424144884"} Mar 16 00:30:05 crc kubenswrapper[4751]: I0316 00:30:05.454856 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a731686455e6c2894d25e2250888e7f028cddef885b53e3d6e210ad424144884" Mar 16 00:30:05 crc kubenswrapper[4751]: I0316 00:30:05.454953 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29560350-dmhzw" Mar 16 00:30:05 crc kubenswrapper[4751]: I0316 00:30:05.863096 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29560344-d5j4b"] Mar 16 00:30:05 crc kubenswrapper[4751]: I0316 00:30:05.873985 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29560344-d5j4b"] Mar 16 00:30:06 crc kubenswrapper[4751]: I0316 00:30:06.464598 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5682a1ec-e3e2-47f2-b820-9b34062bba55" path="/var/lib/kubelet/pods/5682a1ec-e3e2-47f2-b820-9b34062bba55/volumes" Mar 16 00:30:18 crc kubenswrapper[4751]: I0316 00:30:18.123042 4751 scope.go:117] "RemoveContainer" containerID="613123c92a5390f5761e5ba9d788f48b48a01cf99f429953be8f09dbbe984ebb" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515155647472024465 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015155647473017403 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015155644222016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015155644222015463 5ustar corecore